Artificial Intelligence, Natural Language, the Chinese Room, & Reading for Understanding William J. Rapaport Department of Computer Science & Engineering, Department of Philosophy, Department of Linguistics, and Center for Cognitive Science rapaport@buffalo.edu http://www.cse.buffalo.edu/~rapaport What Is AI? 2 Contrasting Views 1. “The science of making machines do things that would require intelligence if done by humans.” – – Marvin Minsky Using humans to tell us how to program computers • • – e.g., play chess, solve calculus problems, … e.g., see, use language, … “classical AI” What Is AI? (continued) 2. “The use of computer programs and programming techniques to cast light on the principles of intelligence in general and human thought in particular.” – – – Margaret Boden Using computers to tell us something about humans. Cognitive science What Is AI? (cont’d) • My view: AI “computational cognition” asks: – How much of cognition is computable? • 2 further questions: – What is computation? • Turing machine – How will we know if (some aspect of) cognition is computable? • Turing test Alan Turing • British, 1912-1954 • “invented” the idea of computation – Turing “machine” • cracked the Nazi “Enigma” code during WW II • devised test for AI: – Turing “test” Turing Machine • Everything that is (humanly) computable can be computed using: – a machine with: • an “infinite” tape divided into squares (with ‘0’ on each) • a movable read-write head – a programming language with only: • 2 nouns: 0, 1 • 2 verbs: move(left or right), print(0 or 1) • 3 grammar rules: – sequence (begin do this; then do that end) » where “this”/“that” = print, move, or any grammatical instruction – selection (if current square=0 then do this else do that) – repetition (while current square=0 [or: current square=1] do this) • Not everything can be computed! – “halting problem”: can’t algorithmically detect infinite loops – can cognition be computed? (How) Can Computers Think? (Human) cognition is computable (?) (Human) cognitive states & processes can be expressed as algorithms They can be implemented on non-human computers How Computers Can Think (cont’d) • Are computers executing such cognitive algorithms merely simulating cognitive states & processes? • Or are they actually exhibiting them? – Do such computers think? • Answer: Turing’s Test Objection: Searle’s Chinese-Room Argument My reply: Computers can understand just by manipulating symbols (like a Turing machine) The Imitation Game MAN WOMAN “I’m the woman” “I’m the woman” INTERROGATOR The Turing Test The Imitation Game COMPUTER MAN WOMAN “I’m the woman” “I’m the woman” INTERROGATOR The Turing Test #2 The Imitation Game COMPUTER MAN MAN WOMAN “I’m the woman” man “I’m the woman” man INTERROGATOR The Turing Test #3 The Imitation Game COMPUTER MAN HUMAN WOMAN “I’m the woman” human “I’m the woman” human INTERROGATOR The Turing Test I Questions Responses H? / C? “I believe that at the end of the century the use of words and general educated opinion will have altered so much that one will be able to speak of machines thinking without expecting to be contradicted.” - Turing 1950 The Chinese-Room Argument • It’s possible to pass TT, yet not (really) think story + questions I H (in Chinese) (who can’t understand Ch.) + responses (Eng.) program (native Chinese speaker) (in fluent Chinese) for manipulating [Ch.] “squiggles” Searle’s Chinese-Room Argument (1) Computer programs just manipulate symbols (2) Understanding has to do with meaning (3) Symbol manipulation alone is not sufficient for meaning (4) No computer program can understand ¬ (3): You can understand just by manipulating symbols! Contextual Vocabulary Acquisition • Could Searle-in-the-room figure out a meaning for an unknown squiggle? • Yes! – Same way you can figure out a meaning for an unfamiliar word from context What Does ‘Brachet’ Mean? (From Malory’s Morte D’Arthur [page # in brackets]) There came a white hart running into the hall with a white brachet next to him, and thirty couples of black hounds came running after them. [66] 2. As the hart went by the sideboard, the white brachet bit him. [66] 3. The knight arose, took up the brachet and rode away with the brachet. [66] 4. A lady came in and cried aloud to King Arthur, “Sire, the brachet is mine”. [66] 10. There was the white brachet which bayed at him fast. [72] 18. The hart lay dead; a brachet was biting on his throat, and other hounds came behind. [86] 1. Where is the “context”? Reader’s Prior Knowledge PK1 PK2 PK3 PK4 Text Reader’s Prior Knowledge PK1 PK2 PK3 PK4 Text T1 Internalized (Co-)Text integrated with Reader’s Prior Knowledge internalization PK1 I(T1) PK2 PK3 PK4 Text T1 Belief-Revised Internalized Text integrated with Reader’s Prior Knowledge internalization PK1 I(T1) PK2 inference PK3 P5 PK4 Text T1 B-R Integrated KB Text internalization PK1 I(T1) PK2 inference PK3 P5 PK4 P6 I(T2) T1 T2 B-R Integrated KB Text internalization PK1 I(T1) PK2 T1 T2 inference PK3 P5 PK4 I(T2) P6 I(T3) T3 B-R Integrated KB Text internalization PK1 I(T1) PK2 T1 T2 inference PK3 P5 PK4 I(T2) P6 I(T3) T3 “Context” for CVA is the reader’s mind, not the (co-)text Reader’s Mind internalization PK1 P7 Text I(T1) PK2 T1 T2 inference PK3 P5 PK4 I(T2) P6 I(T3) T3 • “Words should be thought of not as having intrinsic meaning, but as providing cues to meaning.” – Jeffrey L. Elman, “On the Meaning of Words and Dinosaur Bones: Lexical Knowledge Without a Lexicon” (2009) • “Words might be better understood as operators, entities that operate directly on mental states in what can be formally understood as a dynamical system.” – Jeffrey L. Elman, “On Words and Dinosaur Bones: Where Is Meaning?” (2007) Computational CVA (cont’d) • KB: SNePS representation of reader’s prior knowledge • I/P: SNePS representation of word in its co-text • Processing (“simulates”/“models”/is?! reading): – Uses logical inference, generalized inheritance, belief revision to reason about text integrated with reader’s prior knowledge – N & V definition algorithms deductively search this “belief-revised, integrated” KB (the wide context) for slot fillers for definition frame… • O/P: Definition frame – slots (features): classes, structure, actions, properties, etc. – fillers (values): info gleaned from context (= integrated KB) Cassie learns what “brachet” means: Background info about: harts, animals, King Arthur, etc. No info about: brachets Input: formal-language (SNePS) version of simplified English A hart runs into King Arthur’s hall. • In the story, B12 is a hart. • In the story, B13 is a hall. • In the story, B13 is King Arthur’s. • In the story, B12 runs into B13. A white brachet is next to the hart. • In the story, B14 is a brachet. • In the story, B14 has the property “white”. • Therefore, brachets are physical objects. (deduced while reading; PK: Cassie believes that only physical objects have color) --> (defineNoun "brachet") Definition of brachet: Class Inclusions: phys obj, Possible Properties: white, Possibly Similar Items: animal, mammal, deer, horse, pony, dog, I.e., a brachet is a physical object that can be white and that might be like an animal, mammal, deer, horse, pony, or dog A hart runs into King Arthur’s hall. A white brachet is next to the hart. The brachet bites the hart’s buttock. [PK: Only animals bite] --> (defineNoun "brachet") Definition of brachet: Class Inclusions: animal, Possible Actions: bite buttock, Possible Properties: white, Possibly Similar Items: mammal, pony, A hart runs into King Arthur’s hall. A white brachet is next to the hart. The brachet bites the hart’s buttock. The knight picks up the brachet. The knight carries the brachet. [PK: Only small things can be picked up/carried] --> (defineNoun "brachet") Definition of brachet: Class Inclusions: animal, Possible Actions: bite buttock, Possible Properties: small, white, Possibly Similar Items: mammal, pony, A hart runs into King Arthur’s hall. A white brachet is next to the hart. The brachet bites the hart’s buttock. The knight picks up the brachet. The knight carries the brachet. The lady says that she wants the brachet. [PK: Only valuable things are wanted] --> (defineNoun "brachet") Definition of brachet: Class Inclusions: animal, Possible Actions: bite buttock, Possible Properties: valuable, small, white, Possibly Similar Items: mammal, pony, A hart runs into King Arthur’s hall. A white brachet is next to the hart. The brachet bites the hart’s buttock. The knight picks up the brachet. The knight carries the brachet. The lady says that she wants the brachet. The brachet bays at Sir Tor. [PK: Only hunting dogs bay] --> (defineNoun "brachet") Definition of brachet: Class Inclusions: hound, dog, Possible Actions: bite buttock, bay, hunt, Possible Properties: valuable, small, white, I.e. A brachet is a hound (a kind of dog) that can bite, bay, and hunt, and that may be valuable, small, and white. Algorithms (for Computers) 1. Generate initial hypothesis by “syntactic manipulation” • • Algebra: Solve an equation for unknown value X Syntax: “Solve” a sentence for unknown word X – – 2. Deductively search wide context to update hypothesis • • 3. “A white brachet (X) is next to the hart” X (a brachet) is something that is next to the hart and that can be white. “Define” node/word X in terms of immediately connected nodes/words Look for: class membership, properties, structure, acts, agents, etc. Define” node/word X in terms of some (but not all) other connected nodes/words Output definition “frame” (schema) A Computational Theory of CVA 1. 2. A word does not have a unique meaning. A word does not have a “correct” meaning. Author’s intended meaning for word doesn’t need to be known by reader in order for reader to understand word in context Even familiar/well-known words can acquire new meanings in new contexts. Neologisms are usually learned only from context a) b) c) 3. Every co-text can give some clue to a meaning for a word. • 4. Generate initial hypothesis via syntactic/algebraic manipulation But co-text must be integrated with reader’s prior knowledge Large co-text + large PK more clues Lots of occurrences of word allow asymptotic approach to stable meaning hypothesis a) b) 5. CVA is computable CVA is “open-ended”, hypothesis generation. a) • b) 6. 7. CVA ≠ guess missing word (“cloze”); CVA ≠ word-sense disambiguation Some words are easier to compute meanings for than others (N < V < Adj/Adv) CVA can improve general reading comprehension (through active reasoning) CVA can & should be taught in schools CVA as Symbol Manipulation • We “solved” each sentence for the unknown word – in terms of the rest of the text together with our background knowledge • Just as we can solve an algebra problem – in terms of the rest of the equation • By manipulating symbols – which is what computers do! Computational Natural-Language Understanding • What else is needed? – besides symbol manipulation Mind as a Symbol-Manipulation System To understand language, a cognitive agent must: • Take discourse as input • Understand ungrammatical input • Make inferences & revise beliefs • Make plans – For speech acts – To ask/answer questions – To initiate conversation • Understand plans – Speech-act plans of interlocutor • • • • Construct user model Learn (about world, language) Have background/world/commonsense knowledge Remember – What it heard, learned, inferred, revised = have a mind! • All of this can be done by computers manipulating symbols! Reading for Understanding Research Initiative • Institute of Education Sciences – US Department of Education • 5 or 6 “R&D Core Teams” • 1 or 2 “Network Assessment Teams” • “Manhattan Project” / “Apollo Mission” – to improve (the teaching of) reading for understanding/reading comprehension – $20 million over 5 years A Center for Reading for Understanding • A Research and Development Core Team to Integrate Vocabulary, Writing, Reasoning, Multimodal Literacies, and Oral Discourse to Improve Reading Comprehension • Principal location: UB • Satellite locations: Niagara University, Penn State University • Affiliated school districts: Niagara Falls CSD, Cleveland Hill USD, State College Area SD Projects • Writing Intensive Reading Comprehension – Jim Collins (UB/LAI) • Contextual Vocabulary Acquisition – Bill Rapaport (UB/CSE & CCS) • Multimodal Literacies – Kathleen Collins (Penn State) • Virtual Interactive Environments to improve vocabulary and reading comprehension – Lynn Shanahan & Mary McVee (UB/LAI) • Interactively Modeled Metacognitive Thinking – Rob Erwin (Niagara U) • Bilingualism and Basic Cognitive Processes – Janina Brutt-Griffler (UB/LAI) o Experimental Design & Statistical Analysis – Ariel Aloe (UB/CSEP) Contextual Vocabulary Acquisition — Bill Rapaport (CSE & CCS) 1. Previous research: • 2. A reader's understanding of a word's meaning in a context is a function of both the context (the surrounding words) and the reader's prior knowledge. Already accomplished: a) A procedure for successful CVA can be expressed in terms so precise that they can be programmed into a computer. b) That computational procedure can then be converted into a strategy teachable to human readers 3. We propose to embed this procedure in a curricular intervention that can help readers improve both a) their vocabulary and b) their reading comprehension. CVA (cont’d) • Our goal is not to improve vocabulary per se, but: – to improve reading comprehension by • active thinking about the text • with a specific goal of vocabulary enrichment in mind – and to provide readers with a method that can be used independently… • e.g., when they are reading on their own …to learn new vocabulary and to improve comprehension. CVA (cont’d) • GOFAI: – if we know how to explicitly teach some cognitive task to humans • e.g., play chess, do calculus, prove theorems – then we can explicitly program a computer to do that task pretty much as humans do it. • Our CVA algorithms fall into this category – Can we do the converse? – Can a human reader learn vocabulary & improve reading comprehension by carrying out our algorithm? CVA (cont’d) • Not “think like a computer” – I.e., rigidly, mechanically, uncreatively • But: – what we have learned by teaching a computer to do CVA can now help us teach human readers who need guidance in CVA. CVA (cont’d) • Research questions: 1. Can a computer algorithm be translated into a successful curricular intervention? a) Does this computer-based curriculum improve (meaning) vocabulary? CVA students vs. “typical context-based” students vs. “no treatment” control – Each read same passage with single unfamiliar word & figure out a meaning from context ii. CVA students vs. “direct-method” control at time t – Tested on meaning at time t > t i. CVA (cont’d) 1. Can a computer algorithm be translated into a successful curricular intervention? b) Does the algorithm-based curriculum improve reading comprehension? • • CVA students vs. “typical context-based” students vs. no-treatment students Test for reading comprehension on passages… – containing unfamiliar words – with no unfamiliar word CVA (cont’d) 2. Other issues: • • “teacher training”/professional development oral discussion: – • • use thinksheet as “detective”/“scientist” notebook role & nature of prior knowledge – • • explicit instruction on using language to reason is valuable what is needed, how to identify it, how to elicit it, etc. test at different ages & in STEM vs. ELA develop software for use in classroom – student can ask Cassie how she figured it out