OASIS Mailing List ArchivesView the OASIS mailing list archive below
or browse/search using MarkMail.

 


Help: OASIS Mailing Lists Help | MarkMail Help

topicmaps-comment message

[Date Prev] | [Thread Prev] | [Thread Next] | [Date Next] -- [Date Index] | [Thread Index] | [Elist Home]


Subject: [xtm-wg] A Proposal (was Re: Knowledge management claims re XTM)





On 24 Aug, EXT Hans Holger Rath wrote:
> Hi,
> 
> My 5 cent (of a Euro :^) about this topic:
> 
> Bryan Thompson wrote:
>> Yes, I agree.  However I feel we could address much of this
>> continually re-emergent confusion if the XTM AG made specific
>> recommendations concerning *how* the XTM (or the ISO standard)
could
>> be used, without modification to the specification, to encode
>> knowledge modeling languages.
> 
> I already proposed such an approach a year ago at GCA's
Metastructures 99,
> Montreal. The ISO WG call them TM templates. A template (or
application profile
> or schema) contains all ontology declarations, transitivity
information,
> type hierarchy information, explicit inference rules, and validation
> constraints.
> 
> Last week I presented the final results at GCA's Extreme Markup
> Languages 2000, Montreal. The concept is based on PSIs (Public 
> Subject Identifiers) which have to be publicly registered by ISO or 
> OASIS or whoever. This registration is everything what is needed,
> no standard has to be changed, but TM tools can refer to the PSIs
and
> support their semantic which is clearly defined together with the
> registration. This method allows definiton of various
> application dependent profiles (e.g. knowledge representation,
> subject classification, and other use cases). 
> 
> My paper is in the Extreme ML proceedings and the appropriate
> XTM WG should have access to it. If requested I can upload a 
> PDF to the egroups server.
> 
> Having such technique at hand knowledge representation can be done
> using TMs. Reading Sowa's book will convince you (he is talking
about
> conceptual graphs which could be expressed as TMs without any
problems).
> 
> Regards,
> --Holger
> 

(A lurker emerges ;-)

I promised Michel in Montreal to post my thoughts on this topic, but
rather than doing so in a more formal manner as I had intended, since
there clearly has been much discussion already about this, I will
throw
out a few informal thoughts and see what the rest of you think.

It seems to me that the issue of encoding particular ontologies and
semantics using TMs is not so much whether the standard is expanded
itself to support them, but rather that there be *some* kind of
standardized mechanisms and definitions so that the portability and
long term usability of data and tools is surer. I.e., whether the
TM spec defines e.g. transitivity of associations or whether there
is a standard template (or other mechanism) for doing so, is not so
critical as there actually *being* such a standard mechanism which
most if not all tools and data employ.

One can encode essentially any ontology using TMs, but one cannot
garuntee that anyone else (particularly any system) will understand
the vocabulary or semantics of that ontology. That is the crux of
the problem. How to e.g. define things such as transitivity in
a way that is standard, portable, and valid over the long term.
One can use structured names, roles, facets, etc. to do this, but
any such approach is limited in utility because it is non-standard.

In my opinion, TMs and XTM will not and cannot reach critical mass
in the marketplace without a core set of standardized ontologies
based on a standardized specification mechanism, because without
that, there is no information interchange -- and it seems to me that
information interchange (or more correctly, information integration)
is the very heart of TMs. To move past "toy" systems, we must have
standard ontologies and means to validate instances to ensure
conformity. This is IMO even more critical than having common PSIs.

These core ontologies should not, IMO (and it seems also in many
others),
be on the same level as say a DTD such as DocBook or TEI, but
would have some degree of official endorsement and promotion in
association with the XTM standard. The idea of defining a core
standard, compatible with the ISO TM standard, and then a set of
official templates or extensions included in or referenced directly
by the standard seems a reasonable approach. It allows the core
to remain concise and "pure" while giving official weight to common
ontologies.

That said, let me offer one concern (and a proposed solution) about
defining templates as sets of PSIs, which is similar to problems
associated with RDF Schemas, and that is rigorous syntactic validation
of instances, from the perspective of information interchange and
tool compatibility.

By simply defining PSIs, inference rules, etc. one cannot ensure
completeness of specification in a TM instance at the level of 
resolution that one (or at least I ;-) would like. One can ensure that 
*only* the defined/licensed PSIs are used, or that conflicting
assertions 
are not made, etc. but one cannot, eg. ensure that e.g. *every*
instance 
of a given element type is associated with a particular classification
in a given ontology, and thus, a software application that would
depend
on an exhaustive classification of topics, associations, association
roles,
etc. could not be sure, by validation prior to processing, that the TM
instance is acceptable input; and must then resort to its own
validation,
which is inefficient and costly. It is no different than having to
process
non-validated but merely well formed XML instances, and equally 
unacceptable where high-volume automated processing and/or data
integrity
are concerned. 

I believe the idea of templates is the way to go, but fear the
definition
of yet another schema language, particularly one which limits itself
to the higher semantic and KR level and not addressing the
nuts-n-bolts
issues of day-to-day data interchange. It also means that vocabularies
are 
defined in two places, one defining syntax, such as a DTD, and another
defining auxilliary semantics, either as a TM or some other encoding.

I would like to thus propose that XML Schemas would serve as an
excellent
solution to this problem (and certainly others). The XTM standard
could
define a core XML Schema which defines the basic syntactic encoding of
TMs in general. This core XML Schema could then be extended and
augmented
with additional Schemas which define both general and more specific 
ontologies, including rigorous constraints for obligatory
classifications,
data formats, etc. The mechanisms provided by XML Schemas for
"subclassing"
would enable the core TM specification to remain free of further
semantics,
while providing a framework for defining official supersets which meet
the needs of various communities.

The beauty of this approach is that instances which happen to conform
to a highly specialized schema can still be processed in a meaningful
way by tools which only understand the core TM schema, as the Infoset
will provide information about the complete "inheritance path" of 
any given element, even if the GI differs from that in the core
Schema. 
This maximizes the utility and benefit of general TM tools while
allowing 
for highly specialized processing, possibly as extensions to general
tools 
analogous to the subclassing of XTM schemas.

Granted, the XML Schema standard is not finished, and that may
certainly
be an obstacle in adopting it in some official manner by the XTM
working
group; however, the timing seems workable, if concensus is reached and
there is cooperation between the XTM and XML Schema working groups.

I will leave the issue of what those "core standard ontologies" might
be to
another thread, though I have a number of thoughts about that too, as
I think it is more critical *how* we define and validate against
those standard ontologies in an explicit and reliable fashion than
what those ontologies are (but only just ;-)

Adopting XML Schemas for XTM both allows the core XTM spec to remain
pure of additional semantics while allowing the standardization of
common ontologies, all within the same framework, and with a clear
hierarchy of relations between general and more specialized
ontologies.

Cheers,

Patrick

--

Patrick Stickler
Senior Specialist
Customer Documentation Applications
Nokia Information Management
Nokia Networks

Hatanpaankatu 1A
33101 Tampere Finland

phone +358 3 25 74808
fax   +358 3 25 75700

patrick.stickler@nokia.com




-------------------------- eGroups Sponsor -------------------------~-~>
GET A NEXTCARD VISA, in 30 seconds!  Get rates
of 2.9% Intro or 9.9% Ongoing APR* and no annual fee!
Apply NOW!
http://click.egroups.com/1/7872/4/_/337252/_/967540625/
---------------------------------------------------------------------_->

To Post a message, send it to:   xtm-wg@eGroups.com

To Unsubscribe, send a blank message to: xtm-wg-unsubscribe@eGroups.com



[Date Prev] | [Thread Prev] | [Thread Next] | [Date Next] -- [Date Index] | [Thread Index] | [Elist Home]


Powered by eList eXpress LLC