| 
 Bill, 
  
Although I think concepts are internal (call them instead 
ideas or semantic senses, if you wish), I think they appropriately point to 
things in the world, so the shared semantics has two aspects:sense and 
reference. The latter is the thing in the world; the former is the 
placeholder for that, and indexed by our language constructs. Insofar as we 
share a common notion of the things in the world, we do have a common semantics, 
i.e., via a common reference. One might say in fact that the degree to which our 
thoughts match or map to the things in the world is the degree to which we have 
a common way of thinking about the things of the world. And then, finally, the 
degree to which our terminology and compositions of our terminology align with 
those common thoughts is the degree to which we can communicate with reasonably 
shared semantics. 
  
What I don't understand is how my term can map directly to 
a thing in the world and bypass my thoughts.  
  
Let's 
do an experiment: 
  
Refer 
directly to a specific tree without 1) using language, 2) pointing, or 3) 
thinking about that tree. I'd say you can probably do without (1) or (2) (e.g., most animals), but how can you do 
without (3)? 
  
Can the tree be physically in my head? If you 
say that the term-reference relation is in my head, then I would say that 
term-reference relation is a concept (yes, an entity, class, relation, property, 
instance, logical operatory, if you will -- as a way of characterizing the 
kinds of concepts), a reified representation in my head, and that that 
term-reference relation in fact can dispense with the term, since we think that 
animals can know the world without necessarily having language. Is the tree in 
the head of my dog? I don't think so: I think the dog has an idea about the 
tree.  
  
I still think that one of the causes of our 
dissonance is that we are talking about 1) ontology, 2) logic, and 3) semantics, 
and not keeping these things straight. I would say that we build engineering 
models (call them engineering ontologies) which try to represent the real world. 
However, those engineering models consist of two items: 1) labels and 2) the 
representation of the meaning of those labels where the meaning is expressed as 
formal classes, entities, relations, properties, instances, rules, etc., 
that are supposed to align with what we think is the way the real world is 
and "means". Now, labels are terms, i.e., the names we give to these 
representations. As such they are elements of our language, abstracted or 
idealized. We have other terms we use in ordinary communication that index those 
labels. Both the terms and the labels are vocabulary; their interpretations, 
i.e., the actual formal models (stand-ins or representations for the real world 
things) they map to, and the mappings, are their semantics. Because 
a logic itself is a language, we have another filegree of potential 
dissonance.  
  
I use "concept" not because I am a conceptualist 
(I'm not), but because I think that that notion abstracts over stuff like 
entity, class, property, relation, attribute, instance, rule, etc. 
 
  
Thanks, 
Leo  _____________________________________________ 
 Dr. Leo Obrst       The MITRE Corporation, 
Information Semantics  lobrst@xxxxxxxxx    Center for 
Innovative Computing & Informatics  Voice: 703-983-6770 7515 
Colshire Drive, M/S H305  Fax: 703-983-1379   McLean, VA 
22102-7508, USA    
   
  
  Hey Leo...
  
  
  Your note hits precisely at the issue that I think plagues concept 
  talk.   
  
  
  (1) If concepts are private to bearers, then why pretend to talk about 
  "shared semantics", "shared meanings" and so on, as is common in the 
  literature on ontology in computer science?  For this to work, there 
  would have to be something (anything!) by virtue of which such concepts could 
  be shared.  There are, as you know, theories about how that may come 
  about (I'm thinking of Carnap's private language argument), but nobody's 
  talking about that in computer science ontology. 
  
  
  (2) If concepts are not private, then there must be some nexus that 
  supports the non-private component of them that is shared.  Generally, we 
  of a realist bent take that to be *reality*.  Barry's comment concerning 
  the bio-ontologist who thinks of their computational bio-ontology 
  representing not concepts, but biological reality, comes to mind.  
  If we still want concepts (say to talk about someone's personal concepts) some 
  form of conceptual realism can be employed to relate the 
two.   
  
  
  In either case, I don't think any useful work is done whatsoever by 
  calling the things denoted by linguistic terms in *computational* ontologies 
  "concepts" with no further comment.  We humans (at least those of us who 
  are not concept theorists) seem to resort to to the use of the term "concept" 
  for the same reason that we call something we can't remember the name of a 
  "thingy" or "whatchamacallit" - in this form it's a kind of forgivable 
  intellectual laziness.  That, or we're *really* talking about concepts in 
  which case we have lots of work to do.  Rather, wouldn't it be better - 
  especially if one doesn't care about (philosophically-motivated) ontology - 
  simply to use the more neutral terms of "property", "relation", and "object" 
  that can be taken to correspond to the denotation of relation- (unary and 
  greater-than-unary) and constant-terms in mathematical logic.  Nicola 
  Guarino, and later with Chris Welty, went this direction.  This relates 
  to Welty's comment of yesterday about there being nothing new in computer 
  science ontology -- it's almost as if computer scientists engaged in the 
  "semantic technology" field are afraid to use terms that might make their 
  enterprise seem less sexy and "semantic", so they stick with 
  "concept"  
  
  On Apr 20, 2007, at 21:27 , Obrst, Leo J. wrote: 
  
    [Opinion on] 
      
    Everything is a concept: entities, relations among 
    them, properties, attributes, even many instances/individuals (days of the 
    week; Joe Montana; etc.) Especially when you think of concept in animal 
    mental apparatus as a placeholder for something real in the real world (I am 
    a realist). Sure, I have a concept for 'Joe Montana'. Is that concept a 
    general notion, i.e., a class of something? No.  
      
    The general problem (from my perspective) is that we 
    are typically always addressing two perspectives: 1) ontology, i.e., what 
    exists in the world? and 2) semantics, i.e., what is the relationship 
    between our ways of talking/thinking and those things in the world? To me 
    it's clear that we are talking about (1) things of the world, but our 
    language (and our thought, I would say) interposes another layer or two. I 
    would say there are minimally 3 things: 1) our language (terms and 
    compositions of terms), 2) the senses of terms (and their compositions) 
    which we might characterize as concepts, and 3) real world referents that 
    those senses or concepts somehow point to. In formal semantics, a good 
    theory of reference (i.e., (3)) is hard to come by.  
      
    [Opinion off] 
      
      _____________________________________________ 
     Dr. Leo Obrst       The MITRE 
    Corporation, Information Semantics  lobrst@xxxxxxxxx    Center 
    for Innovative Computing & Informatics  Voice: 703-983-6770 
    7515 Colshire Drive, M/S H305  Fax: 703-983-1379   
    McLean, VA 22102-7508, USA    
       
    
      
      
      me-thinks this is a leftover from DL-speak in which 
      'concept' refers to the classes, not the relationships.  I prefer the 
      broader use of 'concept' whereby one speaks of the concept of having a 
      brother, or of being a mentor (which of course are 
      relationships). 
        
      Good to raise this ambiguity. 
      Mike 
        
        
      ==========================  Michael Uschold  M&CT, Phantom Works  425 373-2845  michael.f.uschold@xxxxxxxxxx  
       ==========================  
      ---------------------------------------------------- 
       COOL TIP: to skip the phone 
      menu tree and get a human on the phone, go to: http://gethuman.com/tips.html 
        
         
      
      Correction.  Second sentence should read: 
      
  
      Are 
      relations not "conceptual" in the way that "concepts" are? 
      
  
      Sorry 'bout that. 
      
      
      On Apr 20, 2007, at 20:57 , Bill Andersen wrote: 
      Pat, 
        
  
        How come "relations" are a separate category from "concepts"?  
        Are relations not "conceptual" in the way that "conceptual" are?  
        If it is the case that 'concept' is just parlor speak for those things 
        that we typically represent with nodes in a taxonomy or unary predicates 
        in a logic, and if 'relation' is used to talk about those things that 
        are not "concepts" (i.e. the things we like to represent with predicate 
        terms of arity greater than one), then the distinction seems 
        artificial.  Should there not be just "concepts" divided into the 
        1-, 2- ... n-ary cases? 
        
  
        .bill 
        
        
        On Apr 20, 2007, at 19:12 , Cassidy, Patrick J. wrote: 
        
          In discussions I use: 
          "A representation of the structure of 
          concepts and the relations 
          between them, in a form that a computer can 
          reason with." 
          
  
          Pat 
          
  
          Patrick Cassidy 
          CNTR-MITRE 
          260 Industrial Way West 
          Eatontown NJ 07724 
          Eatontown: 732-578-6340 
          Cell: 908-565-4053 
          
          
  
          
  
          
            -----Original Message----- 
            
            
            Of Peter F Brown 
            Sent: Friday, April 20, 2007 7:08 PM 
            To: Ontology Summit 2007 Forum 
            Subject: Re: [ontology-summit] Ontology 
            Framework Draft  
            StatementfortheOntology Summit 
            
  
            Too many too's... ;-) 
            
  
            But seriously, are we looking for a Gartner 
            Group-style 4 word  
          mission 
          
            statement to make it sound good, or do we 
            want a formulation that 
            actually does mean something and that we 
            can agree on?  
            Brevity does not 
            always equate with clarity: if I have to 
            choose to sacrifice one, it 
            would be brevity. 
            
  
            Peter 
            
  
            -----Original Message----- 
            
             
          Deborah 
          
            MacPherson 
            Sent: 20 April 2007 16:02 
            To: Ontology Summit 2007 Forum 
            Subject: Re: [ontology-summit] Ontology 
            Framework Draft Statement 
            fortheOntology Summit 
            
  
            
               "a formal description of 
              terms that represent concepts and 
              relationships in as 
              chosen subject matter of 
            interest"  
            
  
            is too long, too much of a mouthful of too 
            many words. 
            
  
            Debbie 
            
  
            
            
              Its almost good enough... But an ontology 
              is more than just about  
            terms. 
            
              
  
              How about: 
              
  
               "a formal description of 
              terms that represent concepts and 
              relationships in as 
              chosen subject matter of interest" 
              
  
              Mike 
              
  
              
  
              
  
              ========================== 
              Michael Uschold 
              M&CT, Phantom Works 
              425 373-2845 
              
              ========================== 
              
  
              ---------------------------------------------------- 
              COOL TIP: to skip the phone menu tree and 
              get a human on   
            the phone, go 
            
              
              
  
              
  
              
  
              -----Original Message----- 
              
              Sent: Friday, April 20, 2007 3:08 
PM 
              To: Ontology Summit 2007 Forum 
              Subject: Re: [ontology-summit] Ontology 
              Framework Draft   
            Statement for 
            
              theOntology Summit 
              
  
              I agree: we've worked with the definition 
              "a formal descriptions of 
              terms and the relationships between them" 
              [1] as being good   
            enough to 
            
              know what we talking about when we're 
              talking about what   
            we're talking 
            
              about...and "good enough" should be good 
              enough. 
              
  
              Peter 
              
  
              [1] From 'OASIS Reference Model for 
              Service-Oriented Architecture',  
            p17, 
            
              see 
              
 
  
            
            
              
  
              
  
              -----Original Message----- 
              
                
          Chris 
          
            
              Welty 
              Sent: 19 April 2007 20:23 
              To: Ontology Summit 2007 Forum 
              Subject: Re: [ontology-summit] Ontology 
              Framework Draft   
            Statement for 
            
              the Ontology Summit 
              
  
              
  
              Surely after 15 years we can do better 
              than "specification of a 
              conceptualization"?  Isn't it time we put 
              that one to rest? 
              
  
              -Chris 
              
  
              Obrst, Leo J. wrote: 
              
                All, 
                
  
                Here is our draft statement about the 
                Ontology Framework.    
            We invite 
            
              
                you to consider and discuss this -- now 
                and in next    
            week's sessions. 
            
              
                We intend this to be an inclusive 
                characterization of what an   
            ontology 
            
              
  
              
                is. Inclusive: meaning that we invite 
                you to consider    
            where you and 
            
              
                your community is with respect to these 
                dimensions. If you have 
                concerns or issues, restatements or 
                elaborations, please    
            let us know 
            
              
                now and next week. This will shortly be 
                posted on the    
            Framework Wiki 
            
              page: 
              
                
 
  
              
 
  
            
            meworksFor 
            
              Consideration. 
              
                
  
                
  
                Thanks much, 
                
  
                Tom Gruber, Michael Gruninger, Pat 
                Hayes, Deborah McGuinness, Leo 
                Obrst 
                
  
                _____________________________________________ 
                Dr. Leo Obrst       The 
                MITRE Corporation, Information Semantics 
                
                Voice: 703-983-6770 7515 Colshire 
                Drive, M/S H305 
                Fax: 703-983-1379   McLean, VA 22102-7508, 
                USA 
                
  
                
  
                
  
                
  
                
 
   
            
 
  
          ---------------------------------------------------------------------- 
          
            
              
                -- 
                
  
                
  
                _________________________________________________________________ 
                
                Subscribe/Config: 
                
                
                Community 
            Files:   
            
            
              
                Community Wiki: 
                
                 
              
  
              -- 
              Dr. Christopher A. Welty          
                        IBM Watson   
            Research Center 
            
              +1.914.784.7055           
                                
              19 Skyline Dr. 
              
              
              
  
              _________________________________________________________________ 
              
              Subscribe/Config: 
              
              
              Community Files:   
            
            
              Community Wiki: 
              
              
              
  
              No virus found in this incoming 
              message. 
              Checked by AVG Free Edition. 
              Version: 7.5.463 / Virus Database: 
              269.5.5/769 - Release Date: 
              19/04/2007 17:56 
              
  
              
  
              No virus found in this outgoing 
              message. 
              Checked by AVG Free Edition. 
              Version: 7.5.463 / Virus Database: 
              269.5.5/769 - Release Date: 
              19/04/2007 17:56 
              
  
              
  
              _________________________________________________________________ 
              
              Subscribe/Config: 
              
              
              Community Files:   
            
            
              Community Wiki: 
              
              
              
  
              _________________________________________________________________ 
              
              Subscribe/Config:  
            
            
              
              Community Files:   
            
            
              Community Wiki:  
            
            
              
              
 
  
            
  
            
  
            --  
            
  
            ************************************************* 
            Deborah L. MacPherson 
            Specifier, WDG Architecture PLLC 
            Projects Director, 
            Accuracy&Aesthetics 
            
  
            The content of this email may contain 
            private 
            and confidential information. Do not 
            forward, 
            copy, share, or otherwise distribute 
            without 
            explicit written permission from all 
            correspondents. 
            
  
            ************************************************** 
            
  
            _________________________________________________________________ 
            
            Subscribe/Config: 
            
            
            Community Files:  
          
          
            Community Wiki: 
            
            
            
  
            _________________________________________________________________ 
            
            Subscribe/Config:  
            
            
            Community Files:  
          
          
            Community Wiki:  
            
            
            
 
  
          
  
          _________________________________________________________________ 
          
          
          
          
          
            
        
        
        Chief Scientist 
        
        3600 O'Donnell Street, Suite 600 
        Baltimore, MD 21224 
        Office: 410-675-1201 
        Cell: 443-858-6444 
   
        
  
        _________________________________________________________________ 
        
        
        
        
        
          
      
      
      Chief Scientist 
      
      3600 O'Donnell Street, Suite 600 
      Baltimore, MD 21224 
      Office: 410-675-1201 
      Cell: 443-858-6444 
    
    
  
    _________________________________________________________________ 
    
    
    
    
    
       
  
  
 |