COGNITIVE SCIENCE 6, 291-330 (1982) Intensi0nal Concepts in Propositional Semantic Networks* A N T H O N Y S. M A I D A Brown University STUART C. S H A P I R O State University of New York at Buffalo An integrated statement is made concerning the semantic status of nodes in a propositional semantic network, claiming that such nodes represent only intensions. Within the network, the only reference to extensionality is via a mechanism to assert that two intensions have the same extension in some world. This framework is employed in three application problems to illustrate the nature of its solutions. The formalism used here utilizes only assertional information and no structural, or definitional, information. This restriction corresponds to many of the psychologically motivated network models. Some of the psychological implications of network processes called node merging and node splitting ore discussed. Additionally, it is pointed out that both our networks and the psychologically based networks are prone to memory confusions about knowing unless augmented by domain-specific inference processes, or by structural information. INTRODUCTION In this paper, we discuss a particular kind o f semantic network. It is a representation o f knowledge consisting o f nodes a n d labeled, directed arcs in which the following conditions hold (cf. Shapiro, 1971): I) each n o d e represents a unique concept; 2) each concept represented in the network is represented by a node; 3) each concept represented in the n e t w o r k is represented by a unique node (the Uniqueness Principle); 4) arcs represent n o n - c o n c e p *The work presented here was partly supported by the National Science Foundation under Grant No. MCS78-02274 and MCS80--06314. We appreciate the comments and suggestions of the members of the Graduate Group in Cognitive Science and of the SNePS Research Group, both of SUNY at Buffalo. We also thank one of the anonymous reviewers for very extensive comments. 291 292 MAIDA AND SHAPIRO tual binary relations between nodes; 5) the knowledge represented about each concept is represented by the structure of the entire network connected to the node representing the concept. The term propositional semantic network is sometimes used to distinguish those semantic networks in which every assertion that can be stored or accessed in the network is considered a concept, and therefore represented by a node, from those networks in which some such assertions are represented by arcs (most notably set membership and subset relationships, cf. Hendrix, 1979, p. 54; or statements without handle nodes, cf. Fahlman, 1979, p. 112). This paper is concerned with propositional semantic networks so understood. Conceivably, conceptual dependency networks (cf. Schank, 1975) could be classified as propositional semantic networks if their syntax were explicated. We are using the term proposition because propositions are the intensions of sentences. The definition also allows for the use of nodes corresponding to functional individuals. We will more closely examine the Uniqueness Principle in order to help understand the semantics of semantic networks. In doing this, we follow the line of research exemplified by Woods (1975) and Brachman (1977, 1979). The major point will be that nodes represent only intensions and not extensions, e.g., individual concepts rather than referents, and propositions rather than truth values. Insisting that semantic networks be allowed to represent only intensions suggests promising approaches to several knowledge representation problems which often lead to confusion. One of our goals is to devise a representation which is rich enough to maintain the subtle distinctions related to referential opacity and extensional equivalence. The purpose of the representation is to provide a substrate by which processes (e.g., programs, production rules, and inference rules) can operate. The first problem we treat concerns indirect reference, originally raised by Frege (1892) and recently discussed by McCarthy (1979). McCarthy has put the problem into the following form: the meaning of the phrase "Mike's telephone number" in the sentence "'Pat knows Mike's telephone number" is the concept of Mike's telephone number, whereas its meaning in the sentence "'Pat dialled Mike's telephone number" is the number itself. Thus if we also have "Mary's telephone number=Mike's telephone number," then "Pat dialled Mary's telephone number" follows, but "'Pat knows Mary's telephone number" does not. (p. 129-130, italics in original) Knowing is said to create an opaque context in its complement position and dialling is said to create a transparent context. The Uniqueness Principle suggests a solution strategy for this problem. We treat the concept of Mike's phone number and the concept of the number itself as distinct intensions, and thereby create a representional substrate with sufficient resolution to control inference processes which differentially apply in opaque and transparent contexts. INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 293 The second problem we treat is that of representing the concept of the truth value of a proposition. Since it is meaningful to talk about the truth value of a proposition independently of whether we know the proposition is true or false, we should have the ability to explicitly represent the truth value if we need to. For example, any semantic representation of the sentence " J o h n knows whether he is taller than Bill" seems to explicitly reference the proposition underlying the sentence " J o h n is taller than Bill." This seems clear after examining the paraphrase " J o h n knows the truth value of the proposition John is taller than Bill." Finally, we show how questions may be represented in propositional semantic networks. By this we mean how to represent the proposition that someone requested certain information of someone else. Such a proposition is contained in the sentence " I got mad because John asked me whether I was a boy or a girl." HISTORY OF THE PROBLEMS In this section, we briefly review the histories of three problems in knowledge representation for which we propose new solutions later in the paper. The main thrust of this paper is an investigation of the semantics of one kind of semantic network. That the theory leads to nice solution strategies for the three problems should be taken as a further explication of and support for the theory. We can trace the emergence in artificial intelligence (AI) of the first problem to a paper by McCarthy and Hayes (1969) and it surfaced in its present form in later papers by McCarthy (1977, 1979). A well known philosophical solution to this problem, offered by Quine (1956) is to treat knows as an operator which creates an opaque context and to disallow substitution of equals by equals in an opaque context. Another approach has been adopted by Moore (1977, 1980). He encodes Hintikka's (1971) modal logic of knowledge within first-order logic and then builds an inference engine for that logic. An approach also using first-order logic is taken by McCarthy (1979) and Creary (1979). These researchers propose to view a concept as an object of discourse within the logic. They have one term which denotes Mike's phone number and another which denotes the concept of that phone number. This enables them to bypass the problem of replacement of equals by equals because the concept of Mike's phone number and the number itself are different entities. We differ from McCarthy in that we use one node to denote the concept of Mike's phone number and another node to denote the concept of the number itself. One attempt to represent information about knowledge and belief in a semantic network was offered by Cohen (1978), but it is unclear how his 294 MAIDA AND SHAPIRO treatment would relate to the issues raised in this paper. His primary goal was to show how a speaker's beliefs about a heater's beliefs influences the speaker's planning of speech acts. Anderson (1977, 1978) has sketched a semantic network based procedure for processing referring expressions. Anderson's approach involves creating two distinct nodes in the network for the two concepts (i.e., the concept of Mike's phone number and its referent), and perhaps was inspired by Woods' (1975) important paper. He does not work out the details of representing belief, as they are not his primary goal. He does some interesting reaction time experiments which partly test the psychological reality of the Anderson-Woods proposal and these will be discussed later in the paper. The solution we propose is in the spirit of Anderson but is more thoroughly articulated in the following ways: 1) we specify exactly how the nodes are related and which features of the representation trigger which kinds of inference processes; 2) we provide the solution with well articulated philosophical underpinnings; 3) we point out that these same philosophical assumptions provide an identical solution to some, at first glance, unrelated problems; 4) Anderson's experiments used stimulus materials which involved only transparent contexts. We suggest that the experimental results will not generalize to opaque contexts; and, 5) We point out that Anderson's model, if it is straight-forwardly extended, predicts some counter intuitive memory confusions. The second problem, of representing the notion of truth value, has not received much attention in the AI literature, probably because the problem domains which have been attacked by AI researchers have allowed domain specific solutions. We feel the problem deserves more attention for two reasons. First, the notion of truth value in general and the notion of the truth value of a specific proposition in particular can be objects of thought themselves. Since propositional semantic networks are supposed to be knowledge representations in which every concept (object of thought) is represented by a node, the notion of truth value should also be so represented. Second, having decided that truth values of propositions might be represented by nodes in a semantic network, we quickly find where such nodes can be useful. For example, utterance (1) (1) John knows whether he is taller than Bill. can be taken as an assertion that mentions the truth value of the proposition that John is taller than Bill without taking a position on whether it is true or false. Thus, in order to represent (1) as a proposition about a truth value, we need to be able to represent a truth value independently of the specific value (true or false) it happens to be. An alternative solution used by Allen (1979) and Cohen and Perrault (1979) involves specifying the meaning of "knowing whether" as a disjunction of correct beliefs. For instance, the disjunction (2) could serve as the representation for "John knows whether P . " INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 295 (2) (P & John believes P) or (not P & John believes not P) We claim that this representation uses the intensions P, John believes P, not P, and John believes not P, but it does not use the intension of the truth value o f P, which is, we believe, the intension o f "whether P . " A simpler example which we shall use throughout this paper is shown in (3) with its disjunctive reading in (4). (3) John holds an opinion about whether he is taller than Bill. (4) John believes he is taller than Bill or John believes he is not taller than Bill. This disjunctive solution does not generalize to some embedding verbs such as " w o n d e r , " which is shown as sentence (5). (5) John wondered whether P. Perhaps (5) could be paraphrased as (6a) and thence, via the disjunctive reading of " k n o w s whether" as (6b). We feel, however, that a better paraphrase is (6c), which mentions the concept o f a truth value explicitly. (6a) John wanted to know whether P. (6b) John wanted to know P or to know not P. (6c) John was curious about the truth value of P. Our only claims are that a node for the concept o f a truth value should explicitly exist in the network and is useful for certain representation tasks. We agree that (3) and (4) are logically equivalent and this can be captured by inference rules o f some kind. However, logical equivalence is not the same as intensional identity. The last of the three problems we discuss, that of representing questions, acquires salience because of our position that propositional semantic networks represent only intensions and the combined facts that: 1) intensions are meant to correspond to individual concepts or propositions, and 2) questions are neither individual concepts nor propositions. Other representational schemes can trivially represent questions by tagging the symbol structure which describes the content of the question as being a question and not an assertion. For instance, Schank (1975) represents yes-no questions in conceptual dependency diagrams by indicating that the mode o f the diagram is a question, rather than an assertion. Wh-questions are indicated by placing the symbol " * ? * " in the question slot. That is, the question " W h e r e is the salt?" would be represented as something like (LOC SALT *?*). A propositional network would interpret the notation (LOC SALT *?*) as a proposition stating that the salt is at the question mark, or a proposition involving a free variable. It happily turns out that by viewing yes-no questions as enquiries about truth values we can immediately represent them. Later in the paper, we attempt to generalize this solution to wh-questions as well. 296 MAIDA AND SHAPIRO THE THEORETICAL FRAMEWORK What Does a Semantic Network Model? The first issue we must be clear about is what we intend a semantic network to model. We first exclude two possibilities. One possibility is the real world. This would require nodes to represent objects in the world (as opposed to individual concepts) and facts about such objects (as opposed to propositions). Although some people might be interested in semantic networks as models of the world, we are not. A second possibility is that a semantic network models a corpus of natural language text, or perhaps that a semantic network is a data structure in which a text is stored and from which pieces of the text can be retrieved easily. In this case, nodes of the network would represent words, lexemes, morphemes, strings, phrases, clauses, sentences, and so forth. In Woods (1975), for example, it is argued that the semantic network representation of "The dog that had rabies bit the m a n " must distinguish between the proposition of the main clause, "The dog bit the m a n " and the proposition of the subordinate clause, "The dog had rabies." Woods proposed this for semantic reasons (p. 62). Had he made the proposal for purely syntactic reasons, then he would have been representing sentences as opposed to meanings. Although we feel that semantic networks can be used to model natural language text (cf. Shapiro & Neal, 1982), this is not the use with which we are concerned in this paper. A third possibility, and the one that we are concerned with, is that a semantic network models the belief structure of a thinking, reasoning, language using being (e.g., a human). In this case, nodes represent the concepts and beliefs such a being would have. The point is that these concepts are intensions rather than extensions. This is not, perhaps, the majority view of researchers in "knowledge representation." In their survey of knowledge representation research, Brachman and Smith (1980) asked researchers, "between what two things do you envisage the 'representation' relationship?" (p. 68). They report, "The one interesting thing to be said in summary, it seems, is that the phrase which we use as a commonplace label of our endeavor--'the representation of knowledge'--is perhaps surprisingly not taken in a particularly literal sense...what was considered to be 'represented', typically, were various kind of object [sic] in the world" (p. 71). Nevertheless, in answer to the question, "Would you characterize your inquiry as primarily an investigation into the structure of language, or more as an investigation into the structure of thought? . . . . The great majority gave their prime allegiance to the study of thought" (p. 71). INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 297 lntensions The term "intension" derives from Frege's (1892) term "sense". He was concerned with the relationship between equality and the meanings of designating expressions in a language. The fact that (7a) is true and (7b) is false illustrates the problem which concerned Frege. (7a) Necessarily, the Morning Star is the Morning Star. (7b) Necessarily, the Morning Star is the Evening Star. Frege took this as evidence that the designating phrases "the Morning Star" and "the Evening Star" do not have the same meaning, even though they denote the same object. If the expressions were equal then (7a) and (7b) should have identical meaning. Frege used the term "sense" of an expression to intuitively correspond to the meaning of that expression, and he used the term "reference" to correspond to the denotation of the expression. Carnap (1947) attempted to formalize Frege's notion of the sense of an expression as a function from possible states of affairs to denotations. The function was called an "intension" and the denotation was called an "extension" (cf. Dowty, Wall, & Peters, 1981, p. 145). The approach was refined through Kripke's (1963) semantics of necessity and Montague's intensional logic (p. 145). When we say that nodes of a semantic network represent intensions as opposed to extensions, we mean sense as opposed to reference. Additionally for the purposes of this paper, we will not view intensions as functions, although it might be helpful to do so in the future. When we say that nodes of a semantic network represent intensions we mean intension as Frege (1892), McCarthy (1979), and Creary (1979) view intension, as opposed to Carnap (1947), Kripke (1963), Montague (cf. Dowry, et al.), or Moore (1980). We take intensions to correspond to concepts, ideas, objects of thought, or things which can be conceived of. The Need for lntensional Representations Woods (1975) appears to have been the first to emphasize that some nodes of a semantic network should represent intensions. One reason for this is to enable the cognitive agent being modeled to conceive of things which do not exist in the actual world such as unicorns and Santa Claus. Although unicorns do not exist, they can be reasoned about. Indeed, the reader can say, as McCawley (1980) points out, how many horns a two-headed unicorn would have. Thus, we should be able to describe within a semantic network any conceivable concept, independently of whether it is realized in the actual world, and we should also be able to describe whether in fact it is realized. 298 MAIDA AND SHAPIRO Returning to the Morning Star--Evening Star example, Woods (1975) concluded that, "there must be two mental entities (concepts, nodes, or whatever) corresponding to the two different intensions, morning star and evening star. There is then an assertion about these two intensional entities that they denote one and the same external object (extension)" (p. 50). Woods continues with the observation that "there must be some nodes in the network which correspond to descriptions of entities rather than entities themselves . . . . We have to decide how to tell the two kinds of nodes a p a r t " (p. 68). Semantic network theorists have not universally agreed with this position. Schubert, Goebel, and Cercone (1979), for instance say, " W e take the position that terms (nodes, subnets) already have both extensions and intensions'" (p. 128, italics in original). Brachman (1977) takes a position on the other side of Woods, stating, "Semantic networks are representations of the intensions of natural language designators'" (p. 139, italics in original). Yet Brachman still allows some extensional information in his networks: " s o m e of the operations in the network scheme are purely intensional, while others are n o t " (p. 150). We want to go even further than Brachman and say that all information in the network is intensional. The only reference to extensions will be propositions stating that two intensions pick out the same extension in some world, such as the proposition that the Morning Star is the Evening Star in the actual world. The Absence of a Need for Extensional Representation If, as Woods pointed out, at least some nodes represent intensions, do any represent extensions? Indeed, should every node be seen as having both an intension and an extension as Schubert et al. claim and as most philosophers usually treat designating expressions? Our answer derives from what we take our networks to model (see above). If a network modeled the real world, then a node would represent (or denote) an extension, but we take our networks to model conceptual belief structures. A node that represents only an intension carries no c o m m i t m e n t that an object realizing the intension exists in the real world. The standard translation of " T h e present king of France is b a l d " into a logical notation seems to require asserting the existence of the present king of France. (EXISTS x) (x is the-present-king-of-France & x is bald). However, this is because of the normal extensional interpretation of statements in standard logic. A constant node in a semantic network is like a Skolem constant derived from a extensionally quantified variable that asserts only the existence of the intension. INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 299 S T R U C T U R A L I M P L I C A T I O N S OF I N T E N S I O N A L REPRESENTATION The Need for Co-referential Propositions If a semantic network has a node for the intension of the Morning Star and a different node for the intension of the Evening Star, what should be done when the assertion is made that the Morning Star is the Evening Star? If the nodes were merged by transferring all of the arcs from one node to the other and eliminating the first, then this would eliminate the distinction between the two concepts and make it impossible to represent the sentence " J o h n did not know that the Morning Star is the Evening Star" differently from the sentence " J o h n did not know that the Morning Star is the Morning S t a r . " The solutiori Woods (1975) proposed is to add a node to the network representing the proposition underlying the sentence " T h e Morning Star is co-referential with the Evening Star in the actual w o r l d . " The co-referential proposition can be used by the system's reasoning processes to infer that certain beliefs about the intension o f the Morning Star can be transferred to (and from) the intension of the Evening Star. This will be discussed further below. Order Dependency The set of nodes existing in a network will depend not only on what information is presented to the network but also on the order o f presentation. We shall call this property order dependency. Consider Russell's (1906) example, "George IV wished to know whether Scott was the author of Waverly" (p. 108). This apparently came about because, even though Scott was well known at the time, Waverly was published anonymously. George the IV had an intension for Scott and an intension for the author of Waverly but did not know whether they were extensionally equivalent. A semantic network simulating George IV, or just recording this fact, would need a different node for each of these intensions. But what does this imply for how we should represent the information that Scott wrote Ivanhoe (assuming that the sentence "Scott wrote lvanhoe" is our first introduction to the novel)? Must we represent this as two propositions, one for asserting the co-referentiality o f the intension for Scott and the intension for the author of lvanhoe, and another for asserting that the author o f lvanhoe wrote Ivanhoe. Our answer is that the cognitive system creates intensions only as needed for storing information about them. A separate concept was needed for the intension of the author of Waverly only because there was thought of that author before an identification was made with any previous intension. Psychological evidence for this analysis has been provided by the work 300 MAIDA AND SHAPIRO o f Anderson and Hastie (1974), Anderson (1977, 1978), and McNabb (1977). This will be discussed in a later section. This theory would predict that a sentence such as " G e o r g e IV thought that the author o f Waverlywas older than Scott," which requires two intensions having the same extension, would be harder to understand by someone whose first introduction to the intension of the author of Waverly was via the statement, "Scott wrote Waverly" than by someone who had already thought about Scott and the author o f Waverly independently. The second person would already have two intensions to use. The first person would at first access the same intension for both Scott and the author o f Waverly, but would then create at least one new concept for the intension o f the author o f Waverly. We call this process splitting and will return to a more detailed description of it later. This phenomenon would explain part o f the cuteness o f the example "Shakespeare's plays weren't written by him, but by someone else of the same n a m e " (Hofstadter, Clossman, & Meredith, 1980). Opacity as the Norm A system that conforms to the Uniqueness Principle does not need the substitutivity of equals for equals as a basic reasoning rule, because no two distinct nodes are equal. Co-referentiality between two nodes must be asserted by a proposition. It requires inference rules to propogate assertions from one node to another node which is co-referential with it. Thus, intensional representation implies that referential opacity is the norm and transparency must be explicity sanctioned by an inference process, unless nodes are " m e r g e d . " Merging will be discussed in a later section. Connections with Reality The main objection to exclusive intensional representation seems to be that if nodes represent only intensions, how could any alleged understanding system so based have any connections with the outside world? To consider this question, we must endow our modeled cognitive agent with sense and effector organs. We will look at a robot system with sight and manipulators. The robot needs a perceptual system in which some node, set of features, and so forth is triggered consistently when a given object is seen. If it is to communicate reasonably with people about perceptual topics, it must be able to make approximately the same perceptual distinctions that we make. These perceptual nodes need not extensionally represent the objects that trigger them. The perceptual nodes can be connected to the semanticconceptual nodes. This allows the robot to " r e c o g n i z e " objects, although it could be fooled. INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 301 The robot also needs effector organs that operate the manipulators consistently, and connections between some semantic-conceptual nodes and the effector organs so that it can operate its manipulators in a manner dictated by its reasoning (decide what to do). Sensors and effectors, by supplying a connection between some node in the semantic network and some object(s) in the actual world, finally provide referents for the node in one particular world. However, these referents are only exemplars of the concept represented by the node. The significance of the node remains its intension. The Meaning of the Nodes There are a number of formalisms compatible with the definition of propositional semantic network presented at the beginning of this paper. The formalism used here contains only propositions and individual concepts. We will not specify a formal semantics for the network structures because the meaning units of the network, the nodes, violate Frege's Principle of Compositionality (cf. Dowty, Wall, & Peters, 1981; p. 42). If a formal language obeys the principle of compositionality, then for any non-atomic expression in the language, there is a set of rules which enable one to determine the expression's meaning from the meaning of the expression's subexpressions. In turn, the meaning of the subexpressions can be determined on the basis of the meaning of the sub-subexpressions, and so on, recursively. There are two properties of a formal semantic system that must hold before one can even think of writing the above mentioned rules. The networks which appear in this paper have neither of those properties. First, in order for the above mentioned recursion to terminate, the language must have atomic expressions, or semantic primitives. However, the only primitives used in the networks are arc labels, which are non-conceputal, serve a syntactic function, and have no meaning. Second, in order for the principle of compositionality to apply at all, the meaning of an expression must not change when it is embedded in another expression. The meaningful expressions in our network, the nodes, get their meaning from the expressions they are embedded in (in addition to getting meaning from their subexpressions) and thereby change their meaning whenever they are embedded in a new expression. Although our formalism does not obey Frege's principle, its characteristics which violate the principle can be found in Quillian's (1968) writings. Quillian, as Brachman (1979) has pointed out, is most known for suggesting that semantic information be stored in a subclass-superclass hierarchy so that properties could be inherited from superclass to subclass. There is, however, a less well known aspect of his writing that is relevant to 302 MAIDA A N D SHAPIRO our discussion. The next two quotations, taken from Quillian (1968), seem to argue for a data structure that violates the principle of compositionality for the two reasons stated above. In the quotations, we will interpret the term " w o r d c o n c e p t " to mean node. First, Quillian argued that there are no primitives: there are no word concepts as such that are "primitive." Everything is simply defined in terms of some ordered configuration of other things in memory. (p. 239) Second, Quillian argues that the meaning of a node is determined by a search process which begins with that node: a word's full concept is defined in the memory model to be all the nodes that can be reached by an exhaustive tracing process, originating at its initial, patriarchal type node... (p. 238, italics in original) This implies that, as additional structure is added to the network, it changes the modeled cognitive agent's understanding of the concepts represented by every node connected to the added structure. Martin (1981) has applied the term d e c o m p o s i t i o n a l to nodes whose meaning is dependent upon what they are constituents of. Because of this extreme form of decompositionality, the network described here will be inclined to m e m o r y confusions precisely because of these decompositional meanings. The mechanics o f these confusions will be described in a later section. Anderson (1978, p. 51, Figure 1) appears to use decompositional meanings. It is an empirical question as to whether humans are inclined to make the above mentioned confusions. It m a y be the case that we will have to augment the network notation so that the meanings are more stable and c o n f o r m to the second prerequisite for compositionality. Woods' (1975) arguments for the need to distinguish between the definitional information associated with a node and the assertional information associated with a node are relevant here. THE REPRESENTATION CONSTRUCTS Here we introduce the network notation used in this paper. It is not the only network notation compatible with semantic networks but uses perhaps the fewest constructs. The network will have two kinds of nodes which we will call base nodes and proposition nodes. The proposition nodes represerit propositions and the base nodes represent individual concepts. All arcs are directional and are either ascending or descending, and come in pairs such that one arc of a pair is descending and the other is ascending. In the diagrams of this paper, only the descending arc o f a pair is shown but there is always an ascending arc (corresponding to the inverse o f the descending arc) which is not shown. A node is a base node if it has no descending arcs emanating from it other than a L E X arc. Otherwise, a node is a proposition node. INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 303 The L E X Arc The purpose o f the LEX arc is to form an access route f r o m an individual concept to a word in a lexical memory. The L E X arc is a device used in this paper only to simplify the presentation. One drawback of the L E X arc is that its function is different than the function o f the other arc.s. This is because the object which it points to is not an intension, but rather an entity in lexical m e m o r y . Using the L E X arc to associate a name, say " J o h n " , with a concept, say the intension of John, does not give the network knowledge (in a conceptual sense) that J o h n ' s name is " J o h n , " and we cannot tell the network that it has the wrong name for the person it thinks is named " J o h n . " A more consistent way to do this would be to have an intension for the word " J o h n " as well as for the person John, and then use a proposition to assert that the intension for " J o h n " is the word for the intension John. Any links to lexical m e m o r y would then emanate from the intension for " J o h n " and not the intension for John. For the purposes of exposition we are representing an approximate distinction between a concept and a word denoting the concept but we would not be able to describe the process of acquiring word meanings with this scheme. So the current scheme is being used only to m a k e the diagrams and presentation simplier. The important things to know about the L E X arc are that: 1) it points from a concept to a word, and 2) the node it emanates from is an individual concept and not a proposition. Other Descending Arcs Other than the LEX arc, descending arcs always point to arguments o f propositions and the nodes from which they descend represent propositions. A proposition node which has no descending arcs coming into it is said to be a non-dominated node and represents a proposition that is believed by the system. It is also necessary to be able to attach erasable, nonconceptual, assertion tags to proposition nodes which are dominated. The purpose of these tags is to indicate that the system believes those propositions. If a dominated proposition does not have a belief tag then the system has no opinion about its truth. The belief set of the semantic network consists exactly of the nodes which are non-dominated or which are tagged as true by an assertion tag. Extensional Equivalence There is a case flame, called the equivalence case frame, which is used to represent extensional equivalence. This case frame is used to assert that two distinct intensions have the same extension in the real world. It does not say 304 MAIDA AND SHAPIRO what the extension is but rather only that they have the same extension. Thus " T h e Morning Star is the Evening Star" is represented in the simplified network structure of Figure 1 and can be paraphrased as " T h e conceptual object denoted by 'Morning Star' is extensionally equivalent to the conceptual object denoted by 'Evening S t a r ' . " The reason the case frame has a special status is that it will interact with propositions involving knowing and believing in ways that other propositions will not interact. Figure 1. A representation (node M3) for the propositional information in the sentence "The Morning Star is the Evening Star." Knowing and Believing We will also use two different know relations. One will be used for knowing a fact as in the sentence " J o h n knows that he is taller than Bill." The other will indicate familiarity with a concrete or abstract object as in knowing a person or game, or as in the sentence " P a t knows Mike's phone n u m b e r . " We will call the sense of knowing a fact " k n o w 1" and the sense of being acquainted with a thing " k n o w 2 " . We justify these relations on the basis of the introspection that knowing an object is different than knowing a fact. Similarly, we will use two different believe relations, called "believe 1" and "believe 2," to correspond to the two different know relations. As stated earlier a propositional semantic network models the belief system of a cognitive agent, which we will call " t h e system." Whenever any of the relations involving knowing or believing appear in the network, they will represent beliefs about beliefs. We shall treat knowl only as correct belief rather than justified, correct belief for the following reason. If a cognitive agent believes that someone knows a facts, then the agent believes: 1) that the someone believes the fact; 2) the fact is true; and 3) that the someone believes the fact for the INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 305 right reasons (as far as the cognitive agent can tell). The third stipulation about "right reasons" is necessary to rule out belief for the wrong reasons such as superstition or guessing. For example, there is at least one person who believes that white can force a win in chess, and there is at least one person who believes that white cannot. Therefore we know there is a person who has a correct opinion about whether white can force a win. Nonetheless, no one currently knows whether white can force a win. Since we are not able to specify what "believing for the right reasons" means, we will discuss knowing only in the sense of correct belief. McCARTHY'S TELEPHONE NUMBER PROBLEM The Main Example We now address ourselves to M c C a r t h y ' s telephone number problem. What follows is our representation, and its rationale, for sentence sequence (8). (8) Pat knows Mike's phone number; Ed dials Mike's phone number; and, Tony correctly believes that Mike's phone number is the same as Mary's phone number. This sentence sequence illustrates the distinctions upon which McCarthy's example focuses. On the basis of (8) the system should conclude that Ed dials Mary's phone number, yet it should not conclude that Pat knows M a r y ' s phone number. Furthermore, it should not conclude that Tony knows either Mike's or M a r y ' s phone number; however, if the system is subsequently told that Tony knows one o f the numbers in particular, it should conclude that Tony also knows the other number. Figure 2 shows the network representation o f the information contained in (8). (Note: Although it does not appear in the figure, assume M5 has been tagged as true by an assertion tag.) Critical to our discussion is an explication of the system's understanding the concept of node M7 of that figure. Node M7 is: Something which Pat knows, and something which Ed dials, and something which is Mike's phone number, and something that is co-referential with Mary's phone number, and something that Tony knows is co-referential with Mary's phone number. In order to explicate the node's full concept, it was necessary to traverse the entire network, and the network is the sum total of the system's beliefs. The assertion that node M7 is co-referent with M a r y ' s phone number is just as much a part its meaning as the assertion that it is Mike's phone number. However, we do not want the system to decide that Pat knows M a r y ' s phone number, so how do we avoid this? This is where extensional equivalence links acquire their special status. If the system believes that X -J X I W X X I&i ,.I _J Ma X IAI .J phone number Figure 2. A representation for the information in the sentence sequence: "Pat knows Mike's phone number; Ed dials Mike's phone number; and, Tony correctly believes that Mike's phone number is the some as Mary's phone number." 306 INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 307 some agent knows some intensional concept, then it will assume that the agent knows all o f the propositions asserted about the intensional concept except those propositions which consist of the extensional equivalence case frame (e.g., M5) or which contain the extensional equivalence case frame (e.g,, M14). So the system will assume that Pat knows M4, M l l , and MI5. Now the system could be mistaken; Pat need not necessarily know M15. This performance characteristic would predict corresponding kinds o f memory confusions in humans. Further discussion of this appears later in the paper in the section titled "Merging and Splitting N o d e s . " Further Examples To further illustrate how we would use the representation, we will look at two more sentences, and then return to the question of transparency verses opacity. Consider (9) below. (9) Mike has two phones. One is 831-1234. The other is 831-4321. Figure 3 shows how we represent it. The next example illustrates how we handle designators which fail to refer, as seen in (10) below. (10) Mike doesn't have a phone. There are two ways to represent this utterance. The first involves the intension of nothing (for a discussion of this intensional object see, Heath, 1967). It is the notion of non-existence. The representation is depicted in Figure 4, and if one o f the non-dominated nodes were submitted to a language generator (node M4 or M5), it might produce the sentence "Mike's phone number is non-existent." The other way employs universal quantification and, as shown in Figure 5, asserts that for all x, x is not Mike's phone number. The quantification arc is taken from the SNePS semantic network formalism (Shapiro, 1979a). Transparent Relations Propogate by Inference We now return to describing how the representation for McCarthy's telephone number problem can support processes which simulate referential transpareocy and opacity. Opacity is the norm because for instance, in the situation of sentence sequence (8), Mike's phone number and Mary's phone number are distinct intensions which are extensionally equivalent; but extensional equivalence is not equality, so we do not encounter the problem o f substitution of equals for equals. Transparency, however, requires some sort o f inference process. What is needed is an inference or production rule which propogates assertions across equivalence arcs, but which has an ante- .~/ X, ,7 & ,, / ",_/ Figure 3. The representation for Mike having two phones. k l; X _J Figure 4. The representation for Mike not having a phone. O~ MIN "('-R ~ --M- A X ->0 a¢ ,¢ X IM .J X IM .J Figure 5. Another representation for Mike not having a phone. 309 310 MAIDA AND SHAPIRO cedent trigger pattern that matches only assertions involving transparent operators. This entails that the system have the conceptual knowledge that the relation dial is transparent but know is not. Such an inference rule would then enable the system to add to its data base that Ed dials Mary's phone number from the information given in (8); yet the system would not add that Pat knows Mary's phone number, because know is not transparent and would not satisfy the trigger pattern of the inference rule. There are situations in which assertions involving opaque operators can also propogate across equivalence paths. Tony in (8) knows that Mike's phone number and Mary's phone number are extensionally equivalent, so any assertion with Tony as the agent and Mike's phone number (node M7) as the object, regardless of whether it involves an opaque operator, should be able to propogate across that equivalence path. In short then, if a cognitive agent knows that two concepts are extensionally equivalent, then for that agent all operators are transparent with respect to that path. The appendix contains examples of inference rules which propogate assertions across an equivalence path. No claim is made that they are complete. REPRESENTING TRUTH VALUE This section makes several points. First, the concept of the truth value of some proposition is intensionally distinct from its particular truth value (true or false) and at times it is necessary to make this distinction explicit in the representation. Second, the same information can be represented by two different configurations of concepts. This illustrates what we mean by order dependency. And third, we contrast the behavior of the system's beliefs about believing with its beliefs about knowing (correct belief). Figure 6 shows the representation for sentence (1 l), with the proviso that the system has already wondered about the truth value of the proposition underlying the sentence "John is taller than Bill." An implication of our discussion of order dependency is that a distinct intension for the truth value of a proposition will be created only if the system wonders about the truth value of that proposition before actually learning its truth value. (ll) It is true that John is taller than Bill. This treatment embodies the claim that the extension of a proposition is its truth value. Node M8 is the intension for the truth value of the proposition underlying the sentence "John is taller than Bill." Node M6 represents the individual concept of truth. When John wonders whether he is taller than Bill as in sentence (5), he is trying to determine the co-reference of node M8. INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 14,1 ,.I 311 14,1 _j Figure 6. The r e p r e s e n t a t i o n f o r the propositional i n f o r m a t i o n contained in the sentence "It is true that John is t a l l e r than Bill." If Tony believes, as in sentence (1), John knows whether John is taller than Bill, then Tony believes John's description of node M8 is complete even though Tony's is not necessarily complete. We emphasize that the configuration of nodes used to represent (11) depends on the order in which the system thinks of them. If someone is directly told that John is taller than Bill and does not have to wonder about it, then the representation would be much simpler, as shown in Figure 7, in which node M4 represents the proposition underlying the sentence " J o h n is taller than Bill." Since (11) is an extraposed version of the sentence "That John is taller than Bill is true," we suspect it carries the presupposition that the listener has in fact wondered about the truth value of the embedded proposition, and so the representation depicted in Figure 6 is the correct one for this sentence. The EQUIV-EQUIV case frame indicates that two intensions already represented in the network are extensionally equivalent in the actual world. Node M6 of Figure 6 would presumably already exist in anyone's memory who has some notion of truth. Mode M8 would exist in people's memory who have wondered about the truth value of "John is taller than Bill" before learning its actual truth value. 312 MAIDA AND SHAPIRO ,.I Figure 7. The representation (node M4) for the propositional information contained in the sentence "John is taller than Bill." The reader might entertain the possibility of using the above technique as a general method o f asserting propositions in the network. This is not possible because it would lead to an infinite regress of assertion embeddings. If the EQUIV-EQUIV case frame were necessary to assert a proposition then Figure 6 would not assert utterance (11) because node M7 has not been asserted by the EQUIV-EQUIV case frame. The convention of taking toplevel nodes as being asserted alleviates this problem. In order to represent sentence (3), duplicated below, we use "believe 2 . " As already mentioned, it means to be familiar (3) John holds an opinion about whether he is taller than Bill. (4) John believes he is taller than Bill or John believes he is not taller than Bill. with, or apprehend, some intension. Figures 8 and 9 contrast our representations o f sentences (3) and (4). In Figure 9, sentence (4) is represented by node M8 which is a proposition involving exclusive-or. A feature inherent in the use of " k n o w 2 " is that the task of appraising in exactly what manner the agent of the " k n o w 2 " relation is familiar with, or apprehends, an individual concept requires the use o f inference rules. " K n o w 2 " says nothing in itself except that the agent is familiar in some domain specific way with the individual concept that is the object of the relation. There must also be a INTENSIONAL CONCEPT IN SEMANTIC NETWORKS 313 believeZ ~J ._1 ._1 ._1 Figure 8. The representation (node M8) for the propositional information contained in the sentence "John holds an opinion about whether he is taller than Bill." way for the specific facts that an agent knows about the concept to be independently asserted. An illustration of this point is depicted in Figures 10 and 11. Node M8 in Figure 10 does not directly represent (12) John knows that he is taller than Bill. sentence (12), but rather entails sentence (12). Node M5 in Figure 11 does directly represent (12). Believing that P and having a correct opinion about :314 MAIDA AND SHAPIRO believe1 REL RE L f...~, ARG2 X Ell r taller Figure 9. The representation (node M8) for the propositional information contained in the sentence "John believes he is taller than Bill or John believes he is not taller than Bill." whether P are two different intensions and therefore should be represented by two distinct nodes. The network structure of Figure 10 does not explicitly assert what John's opinion actually is. Using the inference rule (13) below, the information in Figure 11 could be deduced from the information in Figure 10. INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 315 (13) If the system believesl that some cognitive agent knows2 whether some proposition P and the system believesl that P, then the system can conclude that the agent believesl that P. The following example illustrates the behavior of this kind of system. Suppose you were a college freshman majoring in computer science, and suppose you believed (incorrectly) that pi was a rational number. Naturally, you would also believe that your computer science professor also knew whether pi was rational. By applying the above inference rule, you would erroneously, but properly, conclude that your professor believed that pi was rational. ,.I 144 l¢ kn Jl Figure 10. The representation for the s e n t e n c e s e q u e n c e : "John holds a correct opinion about whether he is taller than Bill; it is true that John is taller than Bill." 316 MAIDA AND SHAPIRO know1 MI Figure 11. The representation (node M5) for the propositional information contained in the sentence "John holds a correct opinion that he is taller than Bill." R E P R E S E N T I N G YES-NO QUESTIONS In order to support the entire range o f discourse processing it is necessary for a knowledge representation in general, and a semantic network in particular, to have the ability to represent questions. Although it is possible for a network which does not represent questions to interface with a processor that enables it to answer questions (e.g., an ATN parser-generator; Shapiro, 1979b), the network itself would intrinsically not have the ability to remember which question were asked. The need to represent questions is illustrated by sentences o f the sort "Since John asked me whether he was taller than Bill, I assumed he had never met Bill." In order to represent this sentence, it is first necessary to represent the question embedded within the subordinate clause. We now turn to the problem of representing a yes-no question in a propositional semantic network. A question is not an individual concept, nor is it a proposition, but it must be represented in a propositional network by some configuration of individual concepts and propositions. It is not the case that a question can be paraphrased into a propositional form. Consider the line of dialog below (14) and the attempt to propositionalize it (15). INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 317 (14) Mary said, "John are you taller than Bill?" (15) Mary asked John whether he was taller than Bill. The paraphrased form superficially resembles a proposition except for a constituent centered around " w h e t h e r , " which cannot be propositionalized. There is a way to circumvent the obstacle by noticing that the sentence resembles sentence (5). We can generate (16) from (15) by replacing " a s k " with " w o n d e r " as shown below. (16) Mary wondered whether John was taller than Bill. Our treatment of " w o n d e r " described earlier applies here and can be extended to " a s k . " Sentence (15) is an enquiry about the truth value of the proposition underlying the sentence " J o h n is taller than Bill." The literal interpretation of all yes-no questions can be handled in this manner. Our representation of (15) is shown in Figure 12. The important thing to note is that the representation captures the fact that the question is an enquiry about the reference o f the truth value of a proposition. The ability to represent this concept supplies the groundwork to build processes (e.g., programs, production rules, inference rules) in the system which uses this concept. The representation can trigger inference rules which will ensure that the system appropriately understands word senses like " a s k . " " A s k " means something like "enquire into the identity o f . " The indirect speech act literature contains many examples o f questions such as, " C a n you pass the salt?" which are not interpreted in natural discourse as questions. Rather they effect the listener as if he had heard some other illocutionary act; in this case it is a request. However, humans are able to understand that, literally, this sentence is a yes-no question and it follows that they must be able to represent this literal interpretation. Furthermore, most accounts of indirect speech acts build on the assumption that the propositional content of the speech act has been extracted. Hence, we would like to represent this literalness in our network. We will discuss wh-questions in the section after next section, after we discuss the process of node splitting. M E R G I N G AND S P L I T T I N G NODES Merging Nodes Many writers point out the need for merging nodes which refer to the same physical object (e.g., Anderson, 1977, 1978; Hofstadter, Clossman, & Meredith, 1980). This kind o f process seems attractive because it would serve to reduce " c l u t t e r " and simplify the crossreferencing problems between information in memory. C o m m o n sense seems to dictate that a system which maintains multiple nodes that are known to be co-referential is inefficieht. This view is expressed by Hofstadter et al.: 318 MAIDA AND SHAPIRO ARG3 ~2 l 144 -J 144 ,.I taller Figure 12. The representation (node M9) for the propositional information contained in the sentence "Mary asks John whether he is taller than Bill." It is inevitable in any representational formalism that distinct nodes will sometimes be manufactured that, it turns out later, stand for the same thing . . . . What happens when one finally finds out that the two nodes represent the same thing? Clearly, they have to be fused somehow into one new node. (p. 4) It is true that this kind o f heuristic leads to increased ability to perform inferences in transparent contexts, however it has the disadvantage o f not being able to inhibit spurious inference processes in domains involving opaque contexts. For instance, as already pointed out, if a system has two separate nodes for Mary's phone number and Mike's phone number, along INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 319 with an assertion that they are extensionally equivalent, then it is not difficult to inhibit the system from concluding that John knows Mike's phone number if it is told that John knows Mary's phone number. But if a merge or copy process were to operate on the two nodes known to be extensionally equivalent, as Hofstadter et al., and Anderson propose, so that all o f the information separately associated with each o f the nodes is merged onto one node, then the spurious inference in the above example would be difficult to inhibit if the system were still going to make transparent inferences. Thus the process of merging nodes must not always take place. Constraints on Merging Nodes One possibility for policing this process is to set the constraint that a maximum of one predicate (assertions about co-reference are exempt) can be asserted per node. This amounts to a total suppression o f any process which merges nodes; and, although it guarantees a representation with sufficient resolution to appropriately trigger both opaque and transparent inference processes, it creates a maximally inefficient system for cross referencing properties among extensionally equivalent concepts in transparent contexts. This option seems drastic and cumbersome, and we concur with Anderson and Hofstadter et al., in that it does not agree with our intuition either. Furthermore, there is experimental evidence that some merging of nodes does take place in humans. Anderson (1977), in addition to demonstrating that a subject can have two distinct nodes which are extensionally equivalent, has obtained evidence that subjects are strongly biased against maintaining two such distinct nodes if they know that the nodes are extensionally equivalent. Usually in such situations one of the two nodes is less strongly memorized than the other node. What subjects are inclined to do during the course of use is to place duplicate copies of the information which reside at the less established node to the more established node and then gradually abandon use o f the less established node. However, back in McCarthy's telephone number problem, the nodes for Mary's phone number and Mike's phone number cannot be merged. The nodes exist in an opaque context. In all o f Anderson's experiments, his stimulus materials were instances o f transparent reference. As we said earlier, in transparent contexts, merging o f nodes leads to optimal performance. Let us look at Anderson's (1978) stimulus materials. Subjects were asked to memorize the following kinds of sentences: (17) (18) (19) (20) (21) The The The The The smart Russian is the tall lawyer. smart Russian cursed the salesgirl. smart Russian rescued the kitten. tall lawyer adopted the child. tall lawyer caused the accident. 320 MAIDA AND SHAPIRO These materials contain no opaque operators. However, we can prefix each o f these study sentences with the phrase " J o h n believes t h a t " ; then if a subject memorizes these modified study sentences, he or she will generate the same subconfiguration of nodes in memory as before but this time within an opaque context. If a subject fully processes the semantics o f the word "believe" (unfortunately, tuning the task demands to force the subject to fully process "believe" might not be so easy) then during the experiment, he or she must process the nodes in a way that maintains the separate intensions. Specifically, the trend in the reaction time data which was observed that indicated subjects were copying information from one coreferent node to another should be absent. We suggest that the process of merging nodes is inhibited when an opaque context is created but not otherwise. This creates a compromise system which will process instances o f transparent reference efficiently yet process instances of opaque reference accurately. Confusions We as yet have not succeeded in devising a representation which does not make any semantic confusions. The problem can be seen by examining Figure 2. After the system is told the information in sentence sequence (8), it will answer " y e s " to the question "Does Pat know Ed dials Mike's phone n u m b e r ? " The reason is that any straight-forward algorithm that adds the assertion to the data base that Pat knows M4 will also cause it to add to the data base that Pat knows M15 because both nodes have the same status in the network. Humans are certainly able to avoid making this mistake, but the representation can be patched up, either by augmenting the processes which use the representation or by augmenting the representation itself. Without changing the representation, the only way to inhibit adding the fact that Pat knows node M15 while still allowing the inference that Pat knows node M4 is by using highly domain specific inference rules. Perhaps humans can generate these hypothesized rules quite quickly. It seems that this approach however would lead to a system whose performance was likely to degrade very time it learned something new because there would be the possibility of a new confusion. With each new confusion, a new domain specific inference rule would have to be created to inhibit the confusion. This is reminiscent of an EPAM-like theory of forgetting. The networks used in this paper use only assertional information. Instead of adding domain specific inference rules, we could augment the representation to include structural, or definitional, information, and then change the semantics of the network so that the meaning o f a node was only what the structural information said what it was. NotationaUy this could be done by adding functional individuals to the present representation con- INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 321 structs (cf. Maida, 1982). The purpose of definitional information would be to define intensions. Although Woods (1975, p. 58) argued for the need to distinguish between structural and assertional information, psychologists have not as yet been inclined to make such a distinction and tend to accept Quillian's formulation (cf. Collins & Loftus, 1975) o f a node's full concept as adequate. Anderson (1976, p. 243) does use structural information to represent transparent versus opaque reference but he does not refer to the distinction in his experimental work (e.g., Anderson, 1978). SpliCing Consider the following situation. What if a memory node created in a transparent context and with several descriptions fused or merged onto it, gets puts into an opaque context? Returning to the phone number problem, what if a cognitive agent learns that Mike and Mary have the same phone number as a consequence of learning that they live together? Then the two descriptions would be attached to the same node as shown in Figure 13. Now suppose this cognitive agent hears sentence (22). (22) Pete doesn't know that Mike's phone number is the same as Mary's phone number. The network structure depicted in Figure 13 could not become part of the representation structure for sentence (22) because that structure treats the concepts of Mike's and Mary's phone numbers as being the same, rather than being extensionaUy equivalent. Our cognitive agent must make a new ) ) Figure 13. The representation for Mike's and Mary's phone number being the same if there is no immediate need to treat them as separate intensions. 322 MAIDA AND SHAPIRO distinction before he can comprehend (22). It seems necessary that two nodes be constructed to separately represent Mike's phone number and Mary's phone number, and they must be made extensionally equivalent to the original node. We call this process splitting and it was mentioned earlier in our discussion of Scott writing Waverly. The final representaiton is shown in Figure 14 in which node M16 represents the proposition of (22). This analysis predicts that a person who learns about Mike's and Mary's phone number in a transparent context will have more difficulty subsequently comprehending (22) than a person who learns this same information in an opaque context because a split must take place. Because of the probable computational overhead involved in splitting nodes, perhaps this kind of strategy might only be utilized when the system is operating in "careful n o d e , " as in social situations where misunderstandings could be embarrassing or costly. Also from a cognitive developmental perspective, this operation should occur late in the developmental sequence. And perhaps the usual mode o f human functioning leads to treating opaque operators as transparent. Witness the difficulty a teacher has in trying to understand why a student does not understand. If the teacher was facile at processing opaque contexts, he or she would be able to accurately assess the student's knowledge state. In light o f the optional nature o f processing associated with opacity, some unanswered experimental questions are raised? What are the factors which trigger opaque processing in humans? Developmentally, when does the ability for opaque processing appear in children? For instance, suppose that a person knows that Jim's wife is Sally's mother and is then told that Mike wants to meet Jim's wife (adapted from Creary, 1979). Under what circumstances might the person assume that Mike realizes he would like to meet Sally's mother and under what circumstances might that person feel the need to determine whether Mike knows Jim's wife is Sally's mother. Consider the following two sentences taken from Anderson (1978). Anderson points out that the former sentence is an instance of (23) I am looking for the best lawyer in town. (24) I am looking for my little old mother. opaque reference and the latter is an instance of transparent reference. Given that the difference in interpretation is probably not attributable to their very slight differing syntactic structure, what is determining whether the interpretation is transparent or opaque? The type of interpretation seems to depend on inference processes taking place in the listener which actively assess the knowledge state of the speaker. To illustrate, consider the interpretation of sentence (25) below, depending on whether a foster child, or an earthquake survivor is speaking. (25) I am looking for my lost mother. a QD 0 Q g- 0 "1 C ~> ..,j C NI LEX IF" LEX LEX ARG1 324 MAIDA AND SHAPIRO In the case of the foster child, the listener realizes that the child has probably never known its mother. Whereas in the case of the earthquake survivor, the listener knows that the survivor has someone specific in mind. REPRESENTING WH-QUESTIONS The treatment of wh-questions becomes clearer when one recognizes the opacity of the complement position of ask. The meaning of sentence (26) is changed if we substitute the phase "Mary's phone number" in place of "Mike's phone number" even if Mike and Mary have the same phone number. (26) Pat asks Joe for Mike's phone number. Therefore, in order to represent (26), we must create a new node which represents the intension of the Mike's phone number which Pat asks Joe for. We then equivalence it to the existing node. The resulting interpretation is depicted in Figure 15. There are additional constraints to consider. We do not want to create a split to answer a simple question if we have already postulated splitting to be computationally expensive; and we want to use as much shared network structure as possible. The representation is constructed and integrated into the rest of memory as follows. The relation ask has three arguments: 1) the person doing the asking; 2) the person asked; and, 3) the thing being asked about. Only the thing being asked about is treated opaquely, so it is the only node that does not get matched to the network as the question is being parsed. Rather, it must be asserted to be extensionally equivalent to some other node in memory. Referring to Figure 15, node M8 gets matched to memory via the network pattern determined by nodes M9, M2, and M1. The node which is found as a result of this pattern match (M7 in this case) will be asserted to be extensionally equivalent to node MS. The system should be able to use information which is asserted about M7 to answer the question. In summary then, the listener constructs a representation of the question, matches it to the rest of the network, and finds a node which plays two roles: 1) it is treated as being extensionally equivalent to the concept being asked about; and, 2) it should serve as the access route to information suitable for answering the question. Note also that the system remembers who asked whom what question. The question is, in fact, an enquiry about a particular intension, and that is the feature the representation unambiguously captures. But it is not so clear for the system as to how to determine what constitutes a good answer to the enquiry about the intension. This sort of task requires domain specific knowledge. It seems necessary to include discourse rules in the )< MI IM JJ ,.I other info X lU ._1 number REL Figure 15. The representation (node MIO) for the propositional information contained in the sentence "Pat asks Joe for Mike's phone number." 325 326 MAIDA AND SHAPIRO system to decide exactly what information the questioner is looking for. It is a mysterious phenomenon that humans perceive wh-questions as asking for specific information; on a literal level, they are vague, (e.g., Where is the salt? In the salt shaker). One possible rule for answering questions about phone numbers is given below in (27). (27) If a person asks you for a description of a phone number that you believe2 or know2, then he wants you to generate a linguistic description of a digit sequence that is extensionally equivalent to the phone number. Lehnert (1978), Allen (1979), and Allen and Perrault (1980) have extensively treated the problem o f deciding what is so specific about specific questions in natural language discourse. A more general rule for answering wh-questions, probably reserved for use when the system cannot find relevant domain specific knowledge, is given below in (28). (28) When answering a wh-question, use any identifying description of the entity that the questioner asked about, but which the questioner himself did not use in formulating the question. Thus if Joe answered (26) by saying, " O h , Mike's phone n u m b e r " then rule (28) would be violated. Alternatively, if he said " O h , that's Mary's phone n u m b e r , " then rule (28) would be satisfied but not rule (27). CONCLUSION This paper makes an integrated statement concerning the semantic status of nodes in a propositional semantic network, claiming that nodes in such a network represent only intensions. Within the network, the only reference to extensionality should be via some mechanism to assert that two intensions have the same extension (e.g., The Morning Star is the Evening Star). We have also shown how processes which simulate referential transparency or referential opacity can operate on the network. Our analyses have been influenced by analytic philosophy, artificial intelligence, and cognitive psychology. We have employed this framework in three application areas to illustrate the nature o f its solutions. First, we map out a solution to McCarthy's telephone number problem, which he devised to succinctly capture the difference between referential transparency and opacity. Second, we directly represent the notion o f the truth value of a proposition as is needed to represent a sentence like " J o h n knows whether he is taller than Bill." Finally, we represent sentences like "Since John asked me whether I was taller than Bill, I assumed he never met Bill." INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 327 We also discuss some of the psychological implications of network processes which we call node merging and node splitting. It has been well recognized that a merge process should occur for co-referential nodes but little attention has been given to restrictions on this merge process and to the need for a splitting process. We theorize that merging should be inhibited in opaque contexts and suggest that Anderson's results on node merging will not generalize to opaque contexts. We also point out the need for a splitting process. The formalism we use in this paper employs only assertional information and no structural, or definitional, information. This is consistent with network models in the psychological literature. In particular, neither Anderson (1978), nor Collins and Loftus (1975) mention the distinction. Our networks our prone to memory confusions to which the psychologically based networks will also be vulnerable. The notation employed in this paper can be augmented, by the use of functional individuals (cf. Maida, 1982) to eliminate the tendency for memory confusions. Whether psychologists should employ the use of functional individuals is an empirical question. This change in formalism, however, results in a drastic change in the semantic interpretation of network structure, as Quillian proposed it. Propositional semantic networks continueto be a fruitful data structure by which to model the organization of the human mind. REFERENCES Allen, J. F. A plan-based approach to speech act recognition (Tech. Rap. 131). Toronto: University of Toronto, Department of Computer Science, 1979. Allen, J. F., & Perrault, C. R. Analyzing intention in utterances. Artificial Intelligence, 1980, /5, 143-178. Anderson, J. R. Language, memory, and/hough/. Hillsdale, N J: Lawrence Erlbaum Associates, 1976. Anderson, J. R. Memory for information about individuals. Memory and Cognition, 1977, 5, 430-442. Anderson, .1. R. The processing of referring expressions within a semantic network. Theore/iva/ is~ues in na/ural language processing-2. New York: Association for Computing Machinery, 1978. Anderson, J. R., & Hastie, R. individuation and reference in memory: Proper names and definite descriptions. Cognitive Ps.vchology, 1974, 6, 495-541. Brachman, R. J. What's in a concept: structural foundations for semantic networks. International Journal f o r Man-Machine Sludies, 1977, 9, 127-152. Brachman, R. J. On the epistemological status of semantic networks. In N. V. Findler (Ed.), A.s:~oc'iative Networks: Represenlalion and use o f knowledge by computers. New York: Academic Press, 1979. Brachman, R. J., & Smith, B. R. Special issue on knowledge representation. S I G A R T N e w s le/ler, February 1980, (70). Carnap, R. Meaning und necessily. Chicago, IL: University of Chicago Press, 1947. 328 MAIDA AND SHAPIRO Cohen, P. R. On knowing what to say: planning speech acts (Tech. Rep. 118). Toronto: University of Toronto, Department of Computer Science, 1978. Cohen, P. R., & Perrault, C. R. Elements of a plan-based theory of speech acts. Cognitive Science, 1979, 3, 177-212. Collins, A. M., & Loftus, E. F. A spreading activation theory of semantic processing. Psychological Review, 1975, 82, 407--428. Creary, L. G. Propositional attitudes: Fregean representation and simulative reasoning. IJCAI Proceedings, 1979, 6, 176-181. Dowty, D. R., Wall, R. E., & Peters, S. Introduction to Montague Semantics. Boston, MA: D. Reidel Publishing Company, 1981. Fahlman, S. NETL: A system f o r representing and using real-world knowledge. Cambridge, MA: MIT Press, 1979. Frege, G. Uber Sinn und Bedeutung. Zeitschrift f ~r Philosophie und philosophiche Kritik, 1892, 100, (all). (Translation available in P. Geach & M. Black (Eds.), Translations from the philosophical writings o f Gottlob Frege. Oxford: Basil Blackwell, 1960. Heath, P. L. Nothing. In P. Edwards (Ed.), The Encyclopedia o f Philosophy (Vol. 5). New York: Macmillan, 1967. Hendrix, G. Encoding knowledge in partitioned networks. In N. Findler (Ed.), Associative Networks: Representation and use o f knowledge by computers. New York: Academic Press, 1979. Hintikka, J. Semantics for propositional attitudes. In L. Linsky (Ed.), Reference and modality. London: Oxford University Press, 1971. Hofstadter, D. R., Clossman, (3. A., & Meredith, M. J. Shakespeare's plays weren't written by him, but by someone else o f the same name (Tech. Rep. 96). Bloomington, IN: Indiana University, Computer Science Department, 1980. Kripke, S. Semantical considerations on modal logic. Acta Philosophica Fennica, 1963, 16, 83-89. Maida, A. S. Using lambda abtraction to encode structural information in semantic networks, 1982, available from the Center for Cognitive Science, Box 1911, Brown University, Providence, RI 02912. Martin, W. A. Roles, co-descriptors, and the formal representation of quantified English expressions. American Journal o f Computational Linguistics, 1981, 7, 137-147. McCarthy, J. Epistemological problems of artificial intelligence. IJCAI Proceedings, 1977, 5, 1038-1044. McCarthy, J. First-order theories of individual concepts and propositions. In J. Hayes, D. Michie, & L. Mikulich (Eds.), Machine Intelligence 9. New York: Halsted Press, 1979. McCarthy, J., & Hayes, P. J. Some philosophical problems from the standpoint of artificial intelligence. In B. Meltzer & D. Michie (Eds.), Machine Intelligence 4. Edinburgh: Edinburgh University Press, 1969. McCawley, J. Paper given at the Symposition in Cognitive Science, Vassar College, Poughkeepsie, NY, April 25-26, 1980. McNabb, S. D. The effects o f encoding strategies and age on the memory representation f o r sentences containing proper names and definite descriptions (Rep. No. 77-3). Bloomington, IN: Indiana University, Indiana Mathematical Psychology Program, 1977. Moore, R. C. Reasoning about knowledge and action. IJCA! Proceedings, 1977, 5, 223-227. Moore, R. C. Reasoning about knowledge and action (Tech. Note 191). SRI International, 1980. Quillian, M. R. Semantic memory. In M. Minsky, (Ed.), Semantic information processing. Cambridge, MA: MIT Press, 1968. Quine, W. V. O. Quantifiers and propositional attitudes. Journal o f Philosophy, 1956, 53, 177-187. INTENSIONAL CONCEPTS IN SEMANTIC NETWORKS 329 Russell, B. On denoting. Mind, 1906, 14. (Reprinted in H. Feigl & W. Sellars (Eds.), Readings in philosophical analysis. New York: Appleton-Century-Crofts, 1949. Schank, R. C. Conceptual information processing. New York: American Elsevier, 1975: Schubert, L. K., Goebel, R. G., & Cercone, N. J. The structure and organization of a semantic net for comprehension and inference. In N. Findler (Ed.), Associative Networks: Representation and use of knowledge by computers. New York: Academic Press, 1979. Shapiro, S. C. A net structure for semantic information storage, deduction and retrieval. IJCAI Proceedings, 1971, 2, 512-523. Shapiro, S. C. Generation as parsing from a network into a linear string. American Journal of Computational Linguistics, 1975, Microfiche 33, 45--62. Shapiro, S. C. Path-based and node-based inference in semantic networks. In D. Waltz (Ed.), Theoretical issues in natural language processing-2. New York: Association for Computing Machinery, 1978. Shapiro, S. C. The SNePS semantic network processing system. In N. Findler (Ed.), Associative Networks: Representation and use of knowledge by computers. New York: Academic Press, 1979. (a) Shapiro, S. C. Generalized augmented transition network grammars for generation from semantic networks. Proceedings of the 17th Annual Meeting of the Association for Computational Linguistics, 1979. (b) Shapiro, S. C., & Neal, J. A knowledge engineering approach to natural language understanding. Proceedings of the 20th Annual Meeting of the Association for Computational Linguistics, 1982. Woods, W. A. What's in a link: the semantics of semantic networks. In D. G. Bobrow & A. M. Collins (Eds.), Representation and Understanding. New York: Academic Press, 1975. APPENDIX This a p p e n d i x c o n t a i n s a list o f inference rules whose p u r p o s e is to p r o p o g a t e assertions across extensional equivalence paths. This is not a c o m p l e t e list to e n a b l e all valid inferences, a n d the specific f o r m a t o f these rules would d e p e n d o n the specific n o t a t i o n o n e used. T h e rules consist o f c o n d i t i o n - a c t i o n pairs in which the c o n d i t i o n specifies a p a t t e r n t h a t must m a t c h the n e t w o r k in o r d e r for the rule to apply a n d the action builds m o r e n e t w o r k structure. T h e c o n d i t i o n o f a rule consists o f a sequence o f clauses which all must m a t c h the n e t w o r k . Clauses specify n e t w o r k structure as follows: (lst-arg-slot relation-slot 2nd-arg slot) or, (property-slot arg-slot) or, (arg-slot equivalent arg-slot) Variables are prefixed with a q u e s t i o n m a r k a n d while u n b o u n d will m a t c h a n y n o d e in the n e t w o r k p r o v i d e d the rest o f the clause m a t c h e s t h a t part o f the n e t w o r k . O n c e a variable matches a node, it is b o u n d to t h a t n o d e for t h e rest o f the application o f that rule. 330 MAIDA AND SHAPIRO Rule 1: (is-transparent ?rel) & (?agent ?rel ?objl) & (?objl equivalent ?obj2) = = > (?agent ?rel ?obj2) Note: For example, If likes is a transparent relation and John likes Mary and Mary is Sally's mother then John likes Sally's mother. However, he may not believe he likes Sally's mother. Rule 2: (?agent believe (?agent ?rel ? o b j l ) ) & (?agent believe (?objl equivalent ?obj2)) - - > (?agent believe (?agent ?rel ?obj2)) If a person believes Jim's wife is Sally's mother and he believes he wants to meet Jim's wife, then he believes he wants to meet Sally's mother regardless of whether Jim's wife in fact is Sally's mother. Rule 3: (?agent know (?agent ?rel ? o b j l ) ) & (?agent believe (?objl equivalent ?obj2)) - - > (?agent believe (?agent ?rel ?obj2)) Note: For example, If John knows he likes Mary and John believes that Mary is Sally's mother then he believes he likes Sally's mother. Although the premises of this rule are stronger than those of rule 2, the conclusion is the same as rule 2. Rule 4: (?agent know (?agent ?rel ? o b j l ) ) & (?agent know (?objl equivalent ?obj2)) - > (?agent know (?agent ?rel ? o b j l ) ) - The conclusion of this rule is stronger than that of rule 3 because the system believes the agent is correct in his belief about the coreferentiality of objects 1 and 2.