[XML-DEV Mailing List Archive Home] [By Thread] [By Date] [Recent Entries] [Reply To This Message] RE: Compressability (WAS RE: XML Schemas: the wrong name)
Hi Martin: I read your document yesterday and what I can discern from one reading is that it is excellent, but like so many meta-level models, hard to apply on the first pass. More examples, please. Topic maps, to me, have always appeared to be very fancy XLink aggregates that are easily implemented as treeviews. In a sense, they do the job that a relational programmer might implement with functions on a treeview object that enable them to query master and child tables JIT. Nodes is nodes. That is not a perjorative comment, but one to point out how something is represented and implemented can diverge as long as the authoritative properties remain invariant. I find I want to compare this to Dr. Newcomb's grove approach. Both approaches lend themselves to the layered architecture and while we may describe layers as above and below, they are actually heterarchies where there are relationships within (domain) and between (ecotonal). >I prefer Thompson's paradigm <-> model <-> classes <-> relations <-> objects ><-> images view. Same here. Clearly, we could model these all declaratively so the connectionist model holds in the abstract. In practice, we need methods to make changes and the API is required. I am still mystified why APIs seem to make the grove designer unhappy. I would suggest (could be dead wrong) that the API is yetAnotherGrove providing the control properties. >From the XML Schema point of view this would seem to equate to >Schema Metamodel <-> schema <-> elements <-> substituion group <-> complex >type <-> simple type >I tried to do the same for ebXML, where I came up with >Business Process Metamodel <-> Business Process <-> Business Document <-> >Functional Unit <-> Aggregate Core Component <-> Basic Core Component >Not absoulutely convinced this is the best representation for the ebXML >material, but it looks like a starting point. Ok so far. In the Hatley-Pirbhai CASE model I adapted for the DTD in Beyond the Book Metaphor, a Process Specification (pspec) metamodel is used to describe a nested hierarchy of processes. However, a Control Specification (cspec) metamodel is also created. If we follow the concept that controls emerge from process relationships, then the model you show for ebXML would be followed by the control specification. When I look at the XLang design, that appears to be a simplified form of the Control Spec. One would route to discoverable interfaces and that routing might be described as the human-designed inter-layer relationships. >> Note the emphasis or recognition of patterns to properly >> name new instances, and the concern of the author that >> most approaches have emphasized relations within a level >> over relations between levels. >What I found particularly interesting is that the relations between levels >are best expressed as parallel operations, whereas the internal ones were >more traditional arcs, which I consider to be serial in nature. Certainly, and that matches the nested hierarchy with controls well. In Beyond the Book, I described the view dimensions to enable the levels of operations or data flow to be linear (binding order dictates each step must be followed in sequence to be well-performed) or non-linear (process can be reordered opportunistically). A view dimension migh better be described now in terms of process encapsulation, but the idea was to infer that at each level or view, the details of a lower levels performance are hidden. This is how a Work Breakdown Structure sees tasks and it enforces the protocol of hierarchy of command. It was also important to describe which events could be viewed between dimensions to enable a control protocol to emerge which remained stable in the face of processes that may be adaptive in process time within a layer. IOW, object encapsulation: I don't care how you did it; I care if what you send is what I asked for, and in some cases where temporal sensitivity is an issue, when I ask for it. >Look at the XML Schema for ISO 13250 Topic Maps that I published last night >at http://www.diffuse.org/TopicMaps/schema.html. This illustrates nicely the >application of the XML Schema version of the model that I gave above. Good reading and I thank you. >David Megginson is convinced that he can do the same with RDF, but I still >question the efficiency with which you can use RDF to describe parallel >relationships (which is what topics actually describe). I leave that to David to defend. :-) I find topic maps easier to understand but that may just be that I've spent more time looking at the independent links of HyTime that became the extended links of XLink. I used the nascent ilink in BTBM but didn't understand it then as well as I might have. It's been ten years, three employers and a lot of committees since then. >>My intuition >> is that just as these authors describe, RDF semantic models >> must be able to be layered and relations between levels >> are critical for compressability to work well. >Agreed. The key thing to look at, as far as I am concerned, is the purpose >of describing the relationships. Precisely. One needs to know the mission of the next layer and in fact, that may be different sets of controls that can access the same resource (which the XLink enables perfectly). >All the RDF examples I've seen are >concerned with assigning metadata to a single resource, not describing the >relationships between sets of resources. In library terms RDF is like >writing out a single catalogue card describing a particular resource. Topic >maps are more like building a subject catalogue from a set of catalogue >cards. Topics identify a set of resources that share common characteristics. Again, we seem to come back to library metaphors. These are ok but perhaps limiting in exploring relationships. When I think of a relationship, I may name it, but I think in terms of an interface. As I told Charles, every place I see that link, I find myself wanting to put a function name there. A topic might be the index on the front of the box holding the cards, but I need to also describe the librarian's functions as well. >> By compressability, >> note the use of automated classification techniques to >> create higher level representations that are more efficient >> and the enabling of higher level operations based on the >> outputs of the lower level processes. >How this can be generalized in XML terms I am at present unclear. XPath >would not seem to help, but it might be possible to do this by means of XSLT >processes associated with XML Schema substitution groups. Unfortunately XSLT >is not really built for parallel processing. Yes, XSLT is what I had in mind too. The literature is clear in citing transformations. However, compressability refers to making a simpler pattern work for the more complex pattern (and in the simple case of downtranslation, it works just fine). The NeoCore guys are claiming XML is perfect for their engine, so maybe more study of their techniques can illuminate this. They scan the data, find the patterns, then iconize them. Is the substitution group a kind of icon? Your description below suggests to me that it is. It becomes a symbol, sort of a hieroglyph, that a trained agent can read. Then perhaps, the iconization is a process by which the pattern is discovered, the icon is created, and an agent is created to handle it. This seems to parallel the process by which one creates data objects and schema, then transforms these to virtual interfaces for which an object is created with methods to handle the data object. Therefore, automated data mining detects a pattern, creates a schema, generates an interface and class for binding data matching the pattern, and iconizes that as a namespace URI. For more complex interlayer relationships, it iconizes as a topic map. That seems to be what SQLServer2000 could provide with OLAP techniques. I am way out on a limb here, so comments from the MS designers would be appreciated. I wonder if OLAP can generate topic maps. >> If we are to use semantic networks well, QOS issues become >> germane quickly. A user of a high-level representation >> should not have to rely on deep knowledge of the network >> to wire it into an advanced workflow. On the other hand, >> unless the QOS numbers indicate a highly reliable component >> operating on an authoritative, credible model, this is a >> dicey thing to try without lots of testing. Exhaustive >> testing seems unreasonable to require, but are there >> alternatives other than "ship it and wait for customer >> feedback". >This is where reusable core components fits into my view of the overall >picture. If we have a reusable abstract component then users can apply this, >or restrictions or extensions thereof, to particular objects (XML elements), >which can then be treated as members of a substituion group that will be >understood by a knowledge agent that is keyed to the relevant type model. I agree. It is only difficult to get folks to adopt pre-existing models and use them. I see three problems everyday: o Managers who refuse to believe that extra-company/industry agreements can be made or will cohere for long enough to be useful. It is the most common objection made to the use of registered definitions. Weirdly, these same people see no problem with the notion that they are wall-to-wall Microsoft shops and are usually doing business with other companies who made the same choice. o Programmers who create their own parsers for XML without understanding that while they have the skill, we end up having to test the heck out of their code at considerable cost and for every version, whereas, if they would use a common service, we have not only our experience, but the aggregate experience of potentially millions of users to verify the component is well-behaved. I understand competitive relationships and tweaking for performance, but unfortunately, the reason is usually ego or unwillingness to understand the component. Human-In-The-Loop. It will be interesting to see if the companies that continue to espouse complexity as a barrier to competition and performance differentiation will be able to stay in business. As someone noted, the web is not about competition first; it is about sustaining alliances, then competition. >This "functional unit" has known properties, but different representations, >and local "tweaks". The user only needs to know the function within a >particular interchange mechanism (business document) for a specific process, >not its internal representation. Which if we go back to the CALS Wars, what was 38784 (the prime TM specification) became 28001 (1000 pages of DTD). It works, but became very hard to apply. The problem is upward aggregation into named monoliths which are then cited without concern for local performance. We certainly need the testing at each level and must be able to decompose and take out a well-encapsulated piece. Both ends of the architecture are problematic. Then we only have to cope with DLLHell. :-) Len Bullard Intergraph Public Safety clbullar@i... http://www.mp3.com/LenBullard Ekam sat.h, Vipraah bahudhaa vadanti. Daamyata. Datta. Dayadhvam.h
|
PURCHASE STYLUS STUDIO ONLINE TODAY!Purchasing Stylus Studio from our online shop is Easy, Secure and Value Priced! Download The World's Best XML IDE!Accelerate XML development with our award-winning XML IDE - Download a free trial today! Subscribe in XML format
|