Well. Just downloaded the Pope’s encyclical on the environment and “our common home”. I’d say it is nothing less than stunning. And in the context of this ontolog discussion, and the challenges that seem to come up here, I wanted to pick up on just one sentence, in this message from Thomas Johnston:
Even when, as you indicate, you are referring to languages for formal ontologies (not to other kinds of formal languages, and not to natural languages), the terminological issues are too vague to pin down. In other words, we mean too many things by "language" and "ontology", i.e. our usages are not well-regulated by agreement on sets of necessary and sufficient conditions for correct usage
There are many themes in this recent conversation here I find fascinating. I spent years exploring things like “the dimensional decomposition of features” and how that relates to classification. And all this kind of exploration does feel critically important to me today, in a global context where “we humans” are having such a hard time bringing ourselves into common focus on high-pressure issues that demand collective response.
So, I have this little phrase that came to me last night, as a way it might be possible to package something that might work.
A ZERO-AMBIGUITY COMMON FOUNDATION
Yes, as Thomas Johnston notes, “the terminological issues are too vague to pin down”. We “mean too many [different] things [by the words we use]”.
Our usages are “are not well-regulated by agreement” – and some will say there’s no way they can be – reality is just too complicated…
So – far too briefly, and no doubt inexpertly, and just as a sketch – I am pushing this thing, as a test/seminal hypothesis, into the maw of confusion.
As suggested a day or two ago in the quote from John Sowa on sets, and the wide range of meanings and interpretations that concept has – and in the context of recognizing that these definitions are “not well-regulated” and maybe cannot be – I am looking for stable bedrock, a non-transient industry standard at the lowest level of all machine-processing languages – which, I am daring to suppose, can probably be translated into an interpretation for natural language as well.
I am looking for a way to stipulate, to intentionally affirm, to postulate, to assert.
For the moment, I am looking at a few basic elements. On the definition of “set” – I like the (?) “absolutely non-confusing” notion that a set is a container – a range of boundaries – maybe “boundary values” – within which something exists. But keep it simple. A set is a container.
So – to make this scientific, and not a matter of opinion or ungrounded floating abstraction, I want to define this container in some hard-core industry-standard way, in a “machine space” – at “the lowest possible level”. Don’t define this thing in terms of “concepts” or ideas or abstractions. Define it in some material/substantial way, in the simplest possible terms that “computer scientists everywhere” might see as basic and obvious.
We got a cell. We got something inside the cell. It’s mechanically represented. It has physical instantiation. Isn’t that what it takes?
Every computer program in the world runs on something like this presumption. Every computer language and program in the world is built up from something like this, in terms of some ascending cascade of abstractions.
Maybe those languages and programs immediately diverge in a zillion ways from this point of departure. But isn’t this the basic foundation for all data assembly, for every structure, for every representation process, for every computation, for every function?
I’ve been gathering just a few basic terms from the conversation here that might be the immediate first elements in some common definition cascade.
For me, it looks like one very robust way to approach this – that does appear to be something like an industry standard – is this general discussion of an “object” – with a very simple model saying something like the entire business of data modeling involves a “real object” and an “abstract object” – and the mapping between them.
And there are some guiding general principles, such as the apparent fact that a taxonomic decomposition is always and necessarily context-specific to some specific environment or motivation. This is the essential core reason we have a zillion overlapping but logically independent semantic ontologies. Every situation is different, no one-size-fits-all, reality never parses the same way twice. But maybe – if we stipulate it in the right way -- all models branch from the same root.
So – we have the “real object” – and we have a locally-motivated context-dependent interpretation of the real object, which enables us to assign which “properties” of this real object belong in our “abstract object” representational model. The entire undertaking of science generally involves this process, and the entire activity of scientific confirmation and validation involves testing the correlation between the abstract object and the real object. “Does this abstract (theoretical) model accurately describe the actual reality?” Test, probe, iterate, confirm….
This is more or less the process in large simple terms.
So, into this context, we propose a few big simple definitions and explorations of “best method” – maybe following some explorations of the meaning of “best” – such as simplest, fastest, most general, etc.
Real object < --------- > Abstract object
Now, if that is right – a few very basic simple definitions – maybe starting with those primal concepts suggested by John –
Similarity, identity, difference
And into that conversation – I would want to suggest that this primal tension between “real object” and “abstract object” has a lot of implications for things we’ve been talking about here on ontolog. For one – all perception of the “real object” involves inference – because we natively see the world of real objects through a lens of presumptive categories. Two “real objects” cannot be defined as identical or not identical – because the properties of “identity” are abstractions and properties of our abstract model, not properties of the “real object” – which actually has no properties innate to itself. Properties are abstract conceptual structures that we insert into our abstract model because they are useful for some context-specific purpose (e.g., this is how this particular medical office or local medical network agrees to handle this particular kind of issue).
Get confused on that point – and we are off the rails. This looks to me like something close to a basic convention.
What is happening for me, I think – is that I am staring at this very fundamental principle (“how to map the real to the abstract”) for the definition of reality, and what it might take to create a universal shared agreement on the semantic fundamentals of abstract description. If that basic process could be mapped in a 100% clear way to the fundamental bedrock of computer processing, where we would probably be safe to assume any language can be constructed, we would have a clear point of focus – a line of demarcation – common ground. Build a zillion processes, in a zillion directions – but build from this common root – where we can always meet each other, and build together in coherent ways --
If that basis were stable – then a bunch of fundamental definitions might quickly emerge – terms like similarity and difference and identity could be given common-ground machine-space (“cellular”) definitions, and a flood of interrelated higher and more composite terms could then also emerge – all kind of terms relating to object description and comparison – terms like analogy or metaphor or “universal” or “class” or “instance” – might then emerge – and maybe not always in the same way, but retaining their context-relative fluency for immediate local applications – such that these terms could continue to be seen as relativistic purpose-specific definitions – as they are in actual practice today.
I can think of a flood of confusing issues that might fall into place if this kind of definition-basis were to emerge in a sound consensual way.
Don’t build all of this on the basis of “natural language” – by empirical observation of natural language -- because natural language is a vastly complex organic and variable thing with no innate perfect rationale. Instead, build it on the basis of a man-made human-constructed intentional and stipulated “synthetic” language – because our synthetic language can be controlled and specified and perfected in micro-detail – and in that context, it is still true that “we can map anything to any desired degree of specificity”, including natural language. There’s nothing we can’t do with it. It’s the ultimate in fine-grained fluency, and comes as close to continuously variable as we feel like pushing it.
An approach like this would give us all a fundamental coherence underlying everything we do. And it would not constrain the endless number of alternative ways people are actually doing things today, because it affirms nothing beyond the absolute foundation. But it would give us all a meeting point for a zillion collaborative developments that the world is calling for today.
Even when, as you indicate, you are referring to languages for formal ontologies (not to other kinds of formal languages, and not to natural languages), the terminological issues are too vague to pin down. In other words, we mean too many things by "language" and "ontology", i.e. our usages are not well-regulated by agreement on sets of necessary and sufficient conditions for correct usage.
But I would suggest, given this proviso, that an ontology is a formalization of the semantic relationships among the members of a lexicon, and a language is a means for combining lexical elements into statements. So it's somewhat analogous to the distinction, in a natural language, between expressions in a language, and the statements created from combinations of those expressions according to grammatical rules.
Of course, the semantic relationships among a set of expressions would themselves be expressed in statements, e.g. the semantic rule that "bachelor" means "unmarried adult male", in an ontology, would be expressed by the statement "A bachelor is an unmarried adult male" in the language for that ontology. So given this fact, a language is indeed a means for expressing the semantic relationships among expressions. The sentences thus produced are analytic sentences, ones true by definition.
But that's not all that a language (for a formal ontology) is. Another class of statements are the synthetic ones, the ones true if and only if what they state to be the case is, in fact, the case. A language for an ontology should be able to express synthetic statements as well.
Since there is no Aristotelian definition (by genus and specific difference) of "ontology" and "ontology language" (within the language community we are both part of), and since, as it seems to me, there will be almost as many putative definitions as there are people attempting to give them, this is little more than my own personal view.
But I would remind anyone else willing to step up to the plate and attempt their own definitions of these terms, that a definition of an _expression_ is more than a collections of true statements about what that _expression_ purportedly represents. As much of an ideal rather than an achievable objective as Aristotelian definition is, it is still the gold standard.