[Home] [By Thread] [By Date] [Recent Entries]
John Cowan wrote: > W. E. Perry wrote: > > > The example I use to illustrate the consequence of this premise is > > that a natively XML database must be able to commit any well-formed > > XML instance--and commit it as an instance of the document type > > described by its root element. It must then be prepared to commit > > *as a further instance of the same document type* any other > > well-formed XML document which exhibits the same root element [...]. > > Why privilege the root element type? Why not require XML databases > to categorize documents based on the presence of U+2022 in the > value of an attribute named "schritschifchisted" in an element > whose grandparent is "yolki-palki"? Because this is XML, and root and document are defined constructs which are required, and which we may therefore confidently expect to find in well-formed instances. [snip] > If you don't know any English, it's no use my writing to you in English, > no matter how carefully I explain myself; and you can't learn English > except from people who already know it. Complete node autonomy is as > useless as completely private languages. But this is not about complete node autonomy! Far from it, in fact. This is about autonomous processes which are dependent upon shared data (in Jon Bosak's felicitous formulation, that data gives them 'something to do'). However, *what* a process does is defined in the first instance by that process, not by the form of the data upon which it operates, or the model or schema to which such data might conform.. The very point of my argument is the question: what is the nature of the 'sharing' of data between otherwise autonomous processes? Where the data conforms to a model or schema or other a priori form, known to all of the processes which might operate upon that data, then we could say that in a very real sense the schema, or model or form (in the Platonic sense) of the data drives the process. That is the model-centric view. I am positing something very different, but which is in my experience much closer to the nature of arbitrary instances of data in the real world. This is that the data instance as originally created has some content and that content is understood by its creator to exhibit certain properties. Some portion of that same content may (and I argue, will) be understood, by processes independent of that original creator, to exhibit certain other properties which makes that data, or content, of interest to those processes precisely because those properties indicate that such data might be put to some use through the operation of those processes. > > it may need to selected for processing based on a process' very > > different understanding of the nature of that content. > > Granted. Where does this understanding come from? The understanding, which is to say the elaboration of particular semantics from the data, comes directly from the operation of the process. That is why only a process can determine the semantics of particular data *in the outcome of the operation of that process, on that occasion and in those particular circumstances*. In practice, this means that if a process can instantiate, from a particular XML instance, data upon which it appears that that process might perform useful work, the only real test of whether it can, in fact, do so is to permit that process to operate and then to put the outcome of that process to use, to see whether it is in fact of the use that had been expected. Respectfully, Walter Perry
|

Cart



