Contextual Vocabulary Acquisition: From Algorithm to Curriculum Michael W. Kibby Department of Learning & Instruction and The Reading Center William J. Rapaport Department of Computer Science & Engineering Department of Philosophy, and Center for Cognitive Science Karen M. Wieland Department of Learning & Instruction , The Reading Center, and The Nichols School NSF ROLE Grant REC-0106338 1 QuickTime™ and a GIF decompressor are needed to see this picture. 2 QuickTime™ and a GIF decompressor are needed to see this picture. QuickTime™ and a GIF decompressor are needed to see this picture. 3 QuickTime™ and a GIF decompressor are needed to see this picture. Why Learning Word Meanings Is Important 4 QuickTime™ and a GIF decompressor are needed to see this picture. Why Learning Word Meanings Is Important Reason 1 National Assessment of Educational ProgressReading (NAEP-Reading) 5 QuickTime™ and a GIF decompressor are needed to see this picture. Meaning Vocabulary Assessment on NAEP-R Meaning vocabulary is the application of one’s understanding of word meanings to passage comprehension. 6 QuickTime™ and a GIF decompressor are needed to see this picture. • Vocabulary knowledge is considered to be one of the five essential components of reading as defined by the No Child Left Behind legislation. 7 QuickTime™ and a GIF decompressor are needed to see this picture. • NAEP will not test definitions in isolation from surrounding text; i.e., – students will not be assessed on their prior knowledge of definitions of words on a list. 8 QuickTime™ and a GIF decompressor are needed to see this picture. Criteria for Selecting Vocabulary Items and Distractors for the NAEP Reading Assessme nt Words Selected for Inclusion Are central to the meaning of the passage, such that lack of understanding may disrupt comprehension May have multiple meanings, but only one meaning will be appropriate in the passage Are likely to be found in grade level reading material Words Excluded from Selection Are technical terms (e.g., photosynthesis, fiduciary) Convey the main idea of the passage (e.g., eminent domain) Are already part of the studentÕ s speaking vocabulary Criteria for Distractors 9 May be other meanings of the target vocabulary word May present other information or content from the text but do NOT present what is meant by the target word May contain other words that look or sound lsimilar to the text word QuickTime™ and a GIF decompressor are needed to see this picture. Examples: • • • • • • • • • • • • • • • • Altruistic Magnanimously Dispersed Impetus Forage Soothing Lost in thought Huddled 10 Abide Piqued Beholden Marathon journey Legacy Abated Social contract Grudge QuickTime™ and a GIF decompressor are needed to see this picture. Why Learning Word Meanings Is Important Reason 2 Learning new things and their words changes or increases our perception and organization of the world 14 QuickTime™ and a GIF decompressor are needed to see this picture. The Lego™ Notion of Learning New Things 15 QuickTime™ and a GIF decompressor are needed to see this picture. Why Learning Word Meanings Is Important Reason 3 • Reading comprehension mandates knowing the meaning (concept, thing) associated with words in the text. • When students do not know meanings of words in a written text, comprehension often disrupted. 16 QuickTime™ and a GIF decompressor are needed to see this picture. 17 QuickTime™ and a GIF decompressor are needed to see this picture. Why Learning Word Meanings Is Important Reason 4 Learning new things and words facilitates students’ abilities to use words judiciously— which is much valued in our society 18 QuickTime™ and a GIF decompressor are needed to see this picture. QuickTime™ and a TIFF (Uncompressed) decompressor are needed to see this picture. 19 QuickTime™ and a GIF decompressor are needed to see this picture. 20 QuickTime™ and a GIF decompressor are needed to see this picture. Why Learning Word Meanings Is Important Reason 5 The Profound Effects of Limited Vocabulary 21 QuickTime™ and a GIF decompressor are needed to see this picture. Profound effects of limited vocabulary continued • Limited vocabulary is associated with lower IQ scores. • Limited vocabulary is associated with limited reading comprehension. – In grades 7+, vocabulary and reading comprehension correlate .75 to .85. 22 QuickTime™ and a GIF decompressor are needed to see this picture. Social Class and Meaning Vocabulary Hart, Betty, & Risley, Todd R. (1995). Meaningful differences in the everyday experience of young children. Baltimore, MD: Brookes. 23 QuickTime™ and a GIF decompressor are needed to see this picture. • Studied 42 children’s vocabulary growth from their 9th month to their 36th month. • Researchers – Visited each child’s home once a month. – Observed and tape recorded for one hour every word spoken to or by child. – Recorded 23-30 hours for every child. 24 QuickTime™ and a GIF decompressor are needed to see this picture. Actual and Estimated Number of Words Heard from 0 - 48 Months 50 In Millions Estimated 40 Working Welfare 30 Actual 20 10 25 Age in Months 48 45 42 39 36 33 30 27 24 21 18 15 12 9 6 3 0 0 Actual & Estimated Words Addressed to Child Professional QuickTime™ and a GIF decompressor are needed to see this picture. “The Invisible Curriculum” 26 QuickTime™ and a GIF decompressor are needed to see this picture. Cumulative Number of New Words (Hart & Risley, 1995) 1200 1000 800 600 400 200 0 10 27 QuickTime™ and a GIF decompressor are needed to see this picture. QuickTime™ and a TIFF (LZW) decompressor are needed to see this picture. 28 QuickTime™ and a GIF decompressor are needed to see this picture. From Algorithm to Curriculum 29 QuickTime™ and a GIF decompressor are needed to see this picture. Definition of “CVA” “Contextual Vocabulary Acquisition” =def • the acquisition of word meanings from text – “incidental” – “deliberate” • by reasoning about – contextual clues – background knowledge (linguistic, factual, commonsense) • Including hypotheses from prior encounters (if any) with the word • without external sources of help – No dictionaries – No people 47 QuickTime™ and a GIF decompressor are needed to see this picture. CVA: From Algorithm to Curriculum 1. Computational theory of CVA – Based on: • • – algorithms developed by Karen Ehrlich (1995) verbal protocols (case studies) Implemented in a semantic-network-based knowledge-representation & reasoning system • SNePS (Stuart C. Shapiro & colleagues) 2. Educational curriculum to teach CVA – – – Based on our algorithms & protocols To improve vocabulary & reading comprehension Joint work with Michael Kibby & Karen Wieland • Center for Literacy & Reading Instruction 48 QuickTime™ and a GIF decompressor are needed to see this picture. People Do “Incidental” CVA • We know more words than explicitly taught – Average high-school grad knows ~45K words learned ~2.5K words/year (over 18 yrs.) – But only taught ~400/school-year • ~ 4800 in 12 years of school (~ 10% of total) Most word meanings learned from context − including oral & perceptual contexts – “incidentally” (unconsciously) • How? 50 QuickTime™ and a GIF decompressor are needed to see this picture. People Also Do “Deliberate” CVA • • • • • • • You’re reading; You understand everything you read, until… You come across a new word Not in dictionary No one to ask So, you try to “figure out” its meaning from “context” How? – guess? derive? infer? deduce? educe? construct? predict? … – our answer: • Compute it from inferential search of “context”, including background knowledge 51 QuickTime™ and a GIF decompressor are needed to see this picture. What does ‘brachet’ mean? 52 QuickTime™ and a GIF decompressor are needed to see this picture. (From Malory’s Morte D’Arthur [page # in brackets]) 1. 2. 3. 4. 10. 18. There came a white hart running into the hall with a white brachet next to him, and thirty couples of black hounds came running after them. [66] As the hart went by the sideboard, the white brachet bit him. [66] The knight arose, took up the brachet and rode away with the brachet. [66] A lady came in and cried aloud to King Arthur, “Sire, the brachet is mine”. [66] There was the white brachet which bayed at him fast. [72] The hart lay dead; a brachet was biting on his throat, and other hounds came behind. [86] 53 QuickTime™ and a GIF decompressor are needed to see this picture. Computational cognitive theory of how to learn word meanings • From context – I.e., text + grammatical info + reader’s prior knowledge • With no external sources (human, on-line) – Unavailable, incomplete, or misleading • Domain-independent – But more prior domain-knowledge yields better definitions • “definition” = hypothesis about word’s meaning – Revisable each time word is seen 54 QuickTime™ and a GIF decompressor are needed to see this picture. Cassie learns what “brachet” means: Background info about: harts, animals, King Arthur, etc. No info about: brachets Input: formal-language (SNePS) version of simplified English A hart runs into King Arthur’s hall. • In the story, B12 is a hart. • In the story, B13 is a hall. • In the story, B13 is King Arthur’s. • In the story, B12 runs into B13. A white brachet is next to the hart. • In the story, B14 is a brachet. • In the story, B14 has the property “white”. • Therefore, brachets are physical objects. (deduced while reading; Cassie believes that only physical objects have color) 55 QuickTime™ and a GIF decompressor are needed to see this picture. --> (defineNoun "brachet") Definition of brachet: Class Inclusions: phys obj, Possible Properties: white, Possibly Similar Items: animal, mammal, deer, horse, pony, dog, I.e., a brachet is a physical object that can be white and that might be like an animal, mammal, deer, horse, pony, or dog 56 QuickTime™ and a GIF decompressor are needed to see this picture. A hart runs into King Arthur’s hall. A white brachet is next to the hart. The brachet bites the hart’s buttock. --> (defineNoun "brachet") Definition of brachet: Class Inclusions: animal, Possible Actions: bite buttock, Possible Properties: white, Possibly Similar Items: mammal, pony, 57 QuickTime™ and a GIF decompressor are needed to see this picture. A hart runs into King Arthur’s hall. A white brachet is next to the hart. The brachet bites the hart’s buttock. The knight picks up the brachet. The knight carries the brachet. --> (defineNoun "brachet") Definition of brachet: Class Inclusions: animal, Possible Actions: bite buttock, Possible Properties: small, white, Possibly Similar Items: mammal, pony, 58 QuickTime™ and a GIF decompressor are needed to see this picture. A hart runs into King Arthur’s hall. A white brachet is next to the hart. The brachet bites the hart’s buttock. The knight picks up the brachet. The knight carries the brachet. The lady says that she wants the brachet. --> (defineNoun "brachet") Definition of brachet: Class Inclusions: animal, Possible Actions: bite buttock, Possible Properties: valuable, small, white, Possibly Similar Items: mammal, pony, 59 QuickTime™ and a GIF decompressor are needed to see this picture. A hart runs into King Arthur’s hall. A white brachet is next to the hart. The brachet bites the hart’s buttock. The knight picks up the brachet. The knight carries the brachet. The lady says that she wants the brachet. The brachet bays at Sir Tor. [background knowledge: only hunting dogs bay] --> (defineNoun "brachet") Definition of brachet: Class Inclusions: hound, dog, Possible Actions: bite buttock, bay, hunt, Possible Properties: valuable, small, white, I.e. A brachet is a hound (a kind of dog) that can bite, bay, and hunt, and that may be valuable, small, and white. 60 QuickTime™ and a GIF decompressor are needed to see this picture. General Comments • System’s behavior human protocols • System’s definition OED’s definition: = A brachet is “a kind of hound which hunts by scent” 61 QuickTime™ and a GIF decompressor are needed to see this picture. Computational cognitive theory of how to learn word meanings from context (cont.) • 3 kinds of vocabulary acquisition: – Construct new definition of unknown word • What does ‘brachet’ mean? – Fully revise definition of misunderstood word • Does “smiting” entail killing? – Expand definition of word used in new sense • Can you “dress” (i.e., clothe) a spear? • Initial hypothesis; Revision(s) upon further encounter(s); Converges to stable, dictionary-like definition; Subject to revision 62 QuickTime™ and a GIF decompressor are needed to see this picture. State of the Art: Computational Linguistics • Information extraction systems • Autonomous intelligent agents • There can be no complete lexicon • Such systems/agents shouldn’t have to stop to ask questions 65 QuickTime™ and a GIF decompressor are needed to see this picture. State of the Art: Computational Linguistics • Granger 1977: “Foul-Up” – Based on Schank’s theory of “scripts” (schema theory) – Our system not restricted to scripts • Zernik 1987: self-extending phrasal lexicon – Uses human informant – Ours system is really “self-extending” • Hastings 1994: “Camille” – Maps unknown word to known concept in ontology – Our system can learn new concepts • Word-Sense Disambiguation: – Given ambiguous word & list of all meanings, determine the “correct” meaning • Multiple-choice test – Our system: given new word, compute its meaning • Essay question 66 QuickTime™ and a GIF decompressor are needed to see this picture. State of the Art: Vocabulary Learning (I) • Elshout-Mohr/van Daalen-Kapteijns 1981,1987: – Application of Winston’s AI “arch” learning theory – (Good) reader’s model of new word = frame • Attribute slots, default values • Revision by updating slots & values – Poor readers update by replacing entire frames – But EM & vDK used: • Made-up words • Carefully constructed contexts – Presented in a specific order 67 QuickTime™ and a GIF decompressor are needed to see this picture. Elshout-Mohr & van Daalen-Kapteijns Experiments with neologisms in 5 artificial contexts • When you are used to a view it is depressing when you live in a room with kolpers. – Superordinate information • At home he had to work by artificial light because of those kolpers. • During a heat wave, people want kolpers, so sun-blind sales increase. – Contexts showing 2 differences from the superordinate • I was afraid the room might have kolpers, but plenty of sunlight came into it. • This house has kolpers all summer until the leaves fall out. – Contexts showing 2 counterexamples due to the 2 differences 68 QuickTime™ and a GIF decompressor are needed to see this picture. State of the Art: Psychology • Johnson-Laird 1987: – Word understanding definition – Definitions aren’t stored – “During the Renaissance, Bernini cast a bronze of a mastiff eating truffles.” 69 QuickTime™ and a GIF decompressor are needed to see this picture. State of the Art: Psychology • Sternberg et al. 1983,1987: – Cues to look for (= slots for frame): • • • • • • • Spatiotemporal cues Value cues Properties Functions Cause/enablement information Class memberships Synonyms/antonyms – To acquire new words from context: • Distinguish relevant/irrelevant information • Selectively combine relevant information • Compare this information with previous beliefs 70 QuickTime™ and a GIF decompressor are needed to see this picture. Sternberg • The couple there on the blind date was not enjoying the festivities in the least. An acapnotic, he disliked her smoking; and when he removed his hat, she, who preferred “ageless” men, eyed his increasing phalacrosis and grimaced. 71 QuickTime™ and a GIF decompressor are needed to see this picture. State of the Art: Vocabulary Learning (II) Some dubious contributions: • Mueser 1984: “Practicing Vocabulary in Context” – • BUT: “context” = definition !! Clarke & Nation 1980: a “strategy” (algorithm?) 1. 2. Look at word & context; determine POS Look at grammatical context • 3. Look at wider context • 4. E.g., “who does what to whom”? [E.g., search for Sternberg-like clues] Guess the word; check your guess 72 QuickTime™ and a GIF decompressor are needed to see this picture. CVA: From Algorithm to Curriculum “guess the word” = “then a miracle occurs” • • 73 Surely, we computer scientists can “be more explicit”! QuickTime™ and a GIF decompressor are needed to see this picture. CVA: From algorithm to curriculum … • Treat “guess” as a procedure call (“subroutine”) – Fill in the details with our algorithm – Convert the algorithm into a curriculum • To enhance students’ abilities to use deliberate CVA strategies – To improve reading comprehension … and back again! • Use knowledge gained from CVA case studies to improve the algorithm • I.e., use Cassie to learn how to teach humans & use humans to learn how to teach Cassie 74 QuickTime™ and a GIF decompressor are needed to see this picture. Why not use a dictionary? Because: • People are lazy (!) • Dictionaries are not always available • Dictionaries are always incomplete • Dictionary definitions are not always useful – ‘chaste’ =df clean, spotless / “new dishes are chaste” – ‘college’ =df a body of clergy living together and supported by a foundation • Most words are learned via incidental CVA, not via dictionaries • Most importantly: – Dictionary definitions are just more contexts! 75 QuickTime™ and a GIF decompressor are needed to see this picture. How Does Our System Work? • Uses a semantic network computer system – semantic networks = “concept maps” – serves as a model of the reader – represents: • reader’s prior knowledge • the text being read – can reason about the text and the reader’s knowledge 82 QuickTime™ and a GIF decompressor are needed to see this picture. Fragment of reader’s prior knowledge: m3 = In “real life”, white is a color m6 = In “real life”, harts are deer m8 = In “real life”, deer are mammals m11 = In “real life”, halls are buildings m12 = In “real life”, b1 is named “King Arthur” m14 = In “real life”, b1 is a king (etc.) 83 QuickTime™ and a GIF decompressor are needed to see this picture. m16 = if & if & if then v3 has property v2 v2 is a color v3 v1 v1 is a kind of physical object 84 QuickTime™ and a GIF decompressor are needed to see this picture. Reading the story: m17 = In the story, b2 is a hart m24 = In the story, the hart runs into b3 (b3 is King Arthur’s hall) – not shown (harts are deer) – not shown 85 QuickTime™ and a GIF decompressor are needed to see this picture. The entire network showing the reader’s mental context consisting of prior knowledge, the story, & inferences. The definition algorithm searches this network & abstracts parts of it to produce a (preliminary) definition of ‘brachet’. 86 QuickTime™ and a GIF decompressor are needed to see this picture. Noun Algorithm Find or infer: • Basic-level class memberships (e.g., “dog”, rather than “animal”) – else most-specific-level class memberships – else names of individuals • • • • • Properties of Ns (else, of individual Ns) Structure of Ns (else …) Functions of Ns (else …) Acts that Ns perform (else …) Agents that perform acts w.r.t. Ns & the acts they perform (else…) • Ownership • Synonyms Else do: “syntactic/algebraic manipulation” • “Al broke a vase” a vase is something Al broke – Or: a vase is a breakable physical object 89 QuickTime™ and a GIF decompressor are needed to see this picture. Verb Algorithm • Find or infer: – Predicate structure: • Categorize arguments/cases – Results of V’ing: • Effects, state changes – Enabling conditions for V • Future work: – Classification of verb-type – Synonyms • [Also: preliminary work on adjective algorithm] 90 QuickTime™ and a GIF decompressor are needed to see this picture. Figure out meaning of word from what? • context (i.e., the text)? – Werner & Kaplan 52, McKeown 85, Schatz & Baldwin 86 • context and reader’s background knowledge? – Granger 77, Sternberg 83, Hastings 94 • context including background knowledge? – Nation & Coady 88, Graesser & Bower 90 • Note: – “context” = text context is external to reader’s mind • Could also be spoken/visual/situative (still external) – “background knowledge”: internal to reader’s mind • What is (or should be) the “context” for CVA? 95 QuickTime™ and a GIF decompressor are needed to see this picture. Some Proposed Preliminary Definitions (to extract order out of confusion) • Unknown word for a reader =def – Word or phrase that reader has never seen before – Or only has vague idea of its meaning • Different levels of knowing meaning of word – Notation: “X” 96 QuickTime™ and a GIF decompressor are needed to see this picture. Proposed preliminary definitions • Text =def – (written) passage – containing X – single phrase or sentence … several paragraphs 97 QuickTime™ and a GIF decompressor are needed to see this picture. Proposed preliminary definitions • Co-text of X in some text =def – The entire text “minus” X; i.e., entire text surrounding X – E.g., if X = ‘brachet’, and text = • “There came a white hart running into the hall with a white brachet next to him, and thirty couples of black hounds came running after them.” Then X’s co-text in this text = • “There came a white hart running into the hall with a white ______ next to him, and thirty couples of black hounds came running after them.” – Cf. “cloze” tests in psychology • But, in CVA, reader seeks meaning or definition – NOT a missing word or synonym: There’s no “correct” answer! – “Co-text” is what many mean by “context” • BUT: they shouldn’t! 98 QuickTime™ and a GIF decompressor are needed to see this picture. Proposed preliminary definitions • The reader’s prior knowledge =def – the knowledge that the reader has when s/he begins to read the text – and is able to recall as needed while reading • “knight picks up & carries brachet” ? small • Warnings: – “knowledge” truth • so, “prior beliefs” is better – “prior” vs. “background” vs. “world”, etc. • See next slide! 99 QuickTime™ and a GIF decompressor are needed to see this picture. Proposed preliminary definitions • Possible synonyms for “prior knowledge”, each with different connotation: – Background knowledge: • Can use for information that author assumes reader to have – World knowledge: • General factual knowledge about things other than the text’s topic – Domain knowledge: • Specialized, subject-specific knowledge about the text’s topic – Commonsense knowledge: • Knowledge “everyone” has – E.g., CYC, “cultural literacy” (Hirsch) • These overlap: – PK should include some CSK, might include some DK – BK might include much DK 100 QuickTime™ and a GIF decompressor are needed to see this picture. Steps towards a Proper Definition of “Context” • Step 1: – The context of X for a reader =def 1. The co-text of X 2. • “+” the reader’s prior knowledge Both are needed! – After reading: • “the white brachet bit the hart in the buttock” most subjects infer that brachets are (probably) animals, from: • • – That text, plus: Available PK premise: “If x bites y, then x is (probably) an animal. Inference is not an enthymeme! • (argument with missing premise) 101 QuickTime™ and a GIF decompressor are needed to see this picture. Proper definition of “context”: • (inference not an enthymeme because): – When you read, you “internalize” the text • • • You “bring it into” your mind Gärdenfors 1997, 1999; Jackendoff 2002 “Missing” premise might be in reader’s mind! – This “internalized text” is more important than the actual words on paper: • • Text: Misread as: “I’m going to put the cat out” “I’m going to put the car out” – leads to different understanding of “the text” – What matters is what the reader thinks the text is, • • Not what the text actually is Therefore … 102 QuickTime™ and a GIF decompressor are needed to see this picture. Proper definition of “context”: • Step 2: – The context of X for a reader =def • A single KB, consisting of: 1. The reader’s internalized co-text of X 2. “+” the reader’s prior knowledge 103 QuickTime™ and a GIF decompressor are needed to see this picture. Proper definition of “context”: • But: What is “+”? – Not: mere conjunction or union! – Active readers make inferences while reading. • From text = “a white brachet” & prior commonsense knowledge = “only physical objects have color”, reader might infer that brachets are physical objects • From “The knight took up the brachet and rode away with the brachet.” & prior commonsense knowledge about size, reader might infer that brachet is small enough to be carried – Whole > sum of parts: • inference from [internalized text + PK] new info not in text or in PK • I.e., you can learn from reading! 104 QuickTime™ and a GIF decompressor are needed to see this picture. Proper definition of “context”: • But: Whole < sum of parts! – Reader can learn that some prior beliefs were mistaken • Or: reader can decide that text is mistaken (less likely) • Reading & CVA need belief revision! • operation “+”: – input: PK & internalized co-text – output: “belief-revised integration” of input, via: • Expansion: – addition of new beliefs from ICT into PK, plus new inferences • Revision: – retraction of inconsistent prior beliefs together with inferences from them • Consolidation: – eliminate further inconsistencies 105 QuickTime™ and a GIF decompressor are needed to see this picture. Prior Knowledge Text PK1 PK2 PK3 PK4 106 QuickTime™ and a GIF decompressor are needed to see this picture. Prior Knowledge Text T1 PK1 PK2 PK3 PK4 107 QuickTime™ and a GIF decompressor are needed to see this picture. Integrated KB Text internalization PK1 T1 I(T1) PK2 PK3 PK4 108 QuickTime™ and a GIF decompressor are needed to see this picture. B-R Integrated KB Text internalization PK1 T1 I(T1) PK2 inference PK3 P5 PK4 109 QuickTime™ and a GIF decompressor are needed to see this picture. B-R Integrated KB Text internalization PK1 I(T1) T1 T2 PK2 inference PK3 P5 PK4 I(T2) P6 110 QuickTime™ and a GIF decompressor are needed to see this picture. B-R Integrated KB Text internalization PK1 I(T1) T1 T2 PK2 inference PK3 T3 P5 PK4 I(T2) P6 I(T3) 111 QuickTime™ and a GIF decompressor are needed to see this picture. B-R Integrated KB Text internalization PK1 I(T1) T1 T2 PK2 inference PK3 T3 P5 PK4 I(T2) P6 I(T3) 112 QuickTime™ and a GIF decompressor are needed to see this picture. Note: All “contextual” reasoning is done in this “context”: B-R Integrated KB internalization PK1 P7 Text I(T1) T1 T2 PK2 inference PK3 T3 P5 PK4 I(T2) P6 I(T3) 113 QuickTime™ and a GIF decompressor are needed to see this picture. Proper definition of “context”: • The context that reader R should use to hypothesize a meaning for R’s internalization of unknown word X as it occurs in text T =def – The belief-revised integration of R’s prior knowledge with R’s internalization of the co-text T–X. 114 QuickTime™ and a GIF decompressor are needed to see this picture. This definition agrees with… • Cognitive-science & reading-theoretic views of text understanding – Schank 1982, Rumelhart 1985, etc. • & AI techniques for text understanding: – Reader’s mind modeled by KB of prior knowledge • Expressed in AI language (for us: SNePS) – Computational cognitive agent reads the text, • “integrating” text info into its KB, and • making inferences & performing belief revision along the way – When asked to define a word, • Agent deductively searches this single, integrated KB for information to fill slots of a definition frame – Agent’s “context” for CVA = this single, integrated KB 116 QuickTime™ and a GIF decompressor are needed to see this picture. Problem in Converting Algorithm into Curriculum • “A knight picks up a brachet and carries it away …” • Cassie: – Has “perfect memory” – Is “perfect reasoner” – Automatically infers that brachet is small • People don’t always realize this: – May need prompting: How big is the brachet? – May need relevant background knowledge – May need help in drawing inferences • Teaching CVA =? teaching general reading comprehension – Vocabulary knowledge correlates with reading comprehension 119 QuickTime™ and a GIF decompressor are needed to see this picture. Research Methodology • AI team: – Develop, implement, & test better computational theories of CVA – Translate into English for use by reading team • Reading team: – Convert algorithms to curriculum – Think-aloud protocols • To gather new data for use by AI team • As curricular technique (case studies) 120 QuickTime™ and a GIF decompressor are needed to see this picture. 5-Minute Intermission 121 QuickTime™ and a GIF decompressor are needed to see this picture. An Analysis of Think-Aloud Protocol of Good Readers Using CVA Strategies During Silent Reading 127 QuickTime™ and a GIF decompressor are needed to see this picture. Limited Data Describing CVA Processes • Ames. • Deighton. • Harmon. 128 QuickTime™ and a GIF decompressor are needed to see this picture. Goals of Our Study • Understand CVA processes. – Improve Cassie (computational model). – Build a more effective CVA curriculum. 129 QuickTime™ and a GIF decompressor are needed to see this picture. Beginning and Ending Research Questions • What text cues are used for CVA? • What cognitive processes are used in CVA, including activation of background information? 130 QuickTime™ and a GIF decompressor are needed to see this picture. Research questions continued • What sense or meaning of an unknown word is gained from CVA? • Do prior textual encounters with a hard word affect CVA in a new text? 131 QuickTime™ and a GIF decompressor are needed to see this picture. Methodology: Hard Words • Identified small set of “hard words.” – Some words from earlier work with Cassie. – Dale & O’Rourke and Carroll, Davis & Richman used as guides. – Some words came from scanning science and current event texts. 132 QuickTime™ and a GIF decompressor are needed to see this picture. Methodology: Text Sets • For each hard word: – identified a set of 7-17 authentic texts, – 1-2 instances of each word in each text. • Sometimes, replaced hard word with a neologism (a non-word): e.g., itresia for estuary 133 QuickTime™ and a GIF decompressor are needed to see this picture. Methodology: Participants • High school students. • Excellent or outstanding readers. • Readers given pre-test. 134 QuickTime™ and a GIF decompressor are needed to see this picture. Methodology: Procedures • Worked 1-1 with researcher. • Told there was an unknown word. • Asked to construct a dictionary-like definition of the highlighted target word. • Read each passage, one at a time. • Researcher explained meaning of other unknown words in text. 137 QuickTime™ and a GIF decompressor are needed to see this picture. Procedures continued • When hard word encountered, reader thought aloud while trying to gain sense of word’s meaning. • Sessions recorded on audio tape. • Sessions transcribed verbatim. 138 QuickTime™ and a GIF decompressor are needed to see this picture. Our Procedures Vary from Normal CVA Processes and Conditions • CVA usually activated by disruption of comprehension / text processing. • Then . . . – Reader notes there is a hard word in the text. – Reader notes that lack of that word’s meaning disrupts comprehension. – Reader decides to figure out meaning of word. 139 QuickTime™ and a GIF decompressor are needed to see this picture. Methodology: Analyses of Verbal Protocol • Analyzed transcribed verbal protocol over and over again. • Began by coding. • Changed coding many times. • Right now, still coding some data, but mostly writing multi-page summaries of each protocol. 140 QuickTime™ and a GIF decompressor are needed to see this picture. Overview of Four Sets of Findings 1. Approaches to CVA. 2. Text cues and use of text cues for CVA. 3. Cognitive processes in CVA. 4. Sense of word meaning from CVA. 141 QuickTime™ and a GIF decompressor are needed to see this picture. Findings–1: Readers’ Approaches to CVA • Most read entire text, then returned to target word to work on its meaning. • Some looked first at the target sentence, then read entire text. • A few stopped reading at the target word or end of sentence to work on word meaning immediately. 142 QuickTime™ and a GIF decompressor are needed to see this picture. Findings–2: Use of Text Cues • • Generally, readers did not initially re-inspect text for cues. Rather, think-alouds usually started with readers hypothesizing a meaning on basis of general passage comprehension and background knowledge. 143 QuickTime™ and a GIF decompressor are needed to see this picture. Use of text cues continued • As readers read further passages and were more confirmed in their hypothesis, many then did reinspect text to find support for their hypothesis. • In this re-inspection, readers generally did not select the sentences we had predicted they would use. 144 QuickTime™ and a GIF decompressor are needed to see this picture. These Findings Lead Us in Three Directions: • To an unanticipated, but logical, conclusion about use of text cues. • To curricular implications. • To change our focus from examining the expected use of context cues to identifying the cognitive processes emerging from the data. 145 QuickTime™ and a GIF decompressor are needed to see this picture. Why Was Rereading for Text Cues Not a Reader’s 1st Step? • Researchers knew text cues that provided useful information, but we knew the hard word’s meaning. • Readers did not know the hard word’s meaning, so the relevant text cues had no particular salience for them. 146 QuickTime™ and a GIF decompressor are needed to see this picture. Our Conclusion Using Context for CVA is Easy When You Already Know Meaning of Word! 147 QuickTime™ and a GIF decompressor are needed to see this picture. Curricular Implications • Teachers should model CVA with words they do not know in texts they have not previously seen. • Students should practice CVA with words they do not know in texts they have not previously seen. 148 QuickTime™ and a GIF decompressor are needed to see this picture. Findings-3: Cognitive Processes in CVA 1. Generated hypothesis / built model. 2. Drew inferences from general passage meaning and prior knowledge. 3. Used language cues. 149 QuickTime™ and a GIF decompressor are needed to see this picture. Cognitive Processes in CVA-continued 4. Used information processing/knowledge acquisition processes (Sternberg, 1987). 5. Used knowledge from prior CVA texts. 6. Closure was rare. 150 QuickTime™ and a GIF decompressor are needed to see this picture. CVA Cognitive Processes–1: Hypothesis / Model Building • All readers hypothesized a meaning of the hard word. 151 QuickTime™ and a GIF decompressor are needed to see this picture. On Further Encounters With Word, Readers . . . • Confirmed hypothesis if congruent with text, usually stating rationale; • Revised hypothesis if not congruent, usually stating rationale; or • If hypothesis not congruent with text, but not enough information in text to revise it, readers questioned hypothesis. 152 QuickTime™ and a GIF decompressor are needed to see this picture. Like Elshout-Mohr & Van DaalenKapteijns’ Good Readers • Our readers generally modified hypotheses in keeping with text. • Our readers appeared to “know” that what they stated was a tentative hypothesis. 153 QuickTime™ and a GIF decompressor are needed to see this picture. Unlike Werner & Kaplan (1952) & Deegan (1995) • It was rare for a reader to force text meaning to fit a previous hypothesis about word meaning. 154 QuickTime™ and a GIF decompressor are needed to see this picture. CVA Cognitive Processes–2: Inferences from General Passage Meaning and Prior Knowledge • Prior knowledge was: – Essential. – Idiosyncratic. – Pervasive. 155 QuickTime™ and a GIF decompressor are needed to see this picture. CVA cognitive processes continued: • Observed Inferencing Strategies – Visualizing. – Clarifying. – Self questioning. – Insight. – Summarizing. – Confirming / confidence. 156 QuickTime™ and a GIF decompressor are needed to see this picture. CVA Cognitive Processes–3: Use of Language Cues • Familiar expressions • Figurative language • Connected series 157 QuickTime™ and a GIF decompressor are needed to see this picture. CVA Cognitive Processes–4: Sternberg’s Information Processing • Separating relevant and irrelevant information. • Combining relevant cues from text. • Comparing new text information with hypothesized meaning and prior knowledge. 158 QuickTime™ and a GIF decompressor are needed to see this picture. CVA Cognitive Processes–5: Knowledge from Prior CVA Texts • Content of previous texts became part of prior knowledge. • Readers used this prior knowledge from previous texts in subsequent texts. 159 QuickTime™ and a GIF decompressor are needed to see this picture. CVA Cognitive Processes–6: Closure was Rare • Tentativeness of hypothesis was generally maintained. • Students rarely felt they had solid knowledge of word’s meaning. 160 QuickTime™ and a GIF decompressor are needed to see this picture. Findings–4: Sense of Intended Word Meaning from CVA • Right and wrong are not appropriate descriptors when contrasting meanings from CVA to intended meaning. • Rational and defensible are better descriptors than right or wrong. • Gradual and cumulative over texts. 161 QuickTime™ and a GIF decompressor are needed to see this picture. Sense of the word continued 1. Don’t Know (dk) any meaning. 2. Not intended–No logical justification for sense of word. 3. Not intended–Reasonable justification for sense of word proffered. 4. Not intended–Based on language patterns, not general text meaning. 162 QuickTime™ and a GIF decompressor are needed to see this picture. Sense of the word continued 5. Vague or partial sense of intended meaning. 6. Approximate sense of intended meaning. 7. Nearly full sense of intended meaning. 8. Full intended meaning. 163 QuickTime™ and a GIF decompressor are needed to see this picture. Protocol Study Limitations • We required reader to think about the unknown word and its meaning. – Readers ordinarily might choose to skip or ignore word. – Readers ordinarily might not note word. 164 QuickTime™ and a GIF decompressor are needed to see this picture. Protocol limitations continued • Our readers encountered word in multiple, consecutive texts. – Therefore, readers had immediate memory of the previous encounter. – This is atypical, as readers ordinarily might not encounter new word a second time for a long period. 165 QuickTime™ and a GIF decompressor are needed to see this picture. Protocol limitations continued • We sometimes ask leading questions: e.g., – To activate background knowledge. – To direct reader back to text. – To elicit reasoning processes. – To ask reader to think again. 166 QuickTime™ and a GIF decompressor are needed to see this picture. Protocol limitations continued • At times, we used non-words in place of real words: e.g., schmalion for tatterdemalion vedosarn for taciturn itresia for estuary. 167 QuickTime™ and a GIF decompressor are needed to see this picture. Protocol Study Strengths • Used authentic texts. • Words were, generally, difficult and not previously known by readers. • Used one hard word in repeated texts, not multiple hard words each in a different text. 168 QuickTime™ and a GIF decompressor are needed to see this picture. How to Improve CVA in Classrooms Guess? Magical Mathematical Formula? CVA Strategies! 169 QuickTime™ and a GIF decompressor are needed to see this picture. How to improve CVA continued 1. CVA curriculum should help students learn to apply reasoning and prior knowledge to textual cues. a. Read passage for full comprehension. b. Draw inferences from language, meaning, and prior knowledge. c. Summarize into a meaning-hypothesis frame. 170 QuickTime™ and a GIF decompressor are needed to see this picture. How to improve CVA continued 2. Teachers: a. Model CVA with words whose meanings are unknown. b. Scaffold groups applying CVA strategies with words whose meanings are unknown. c. Guide small groups in think-alouds of CVA. 171 QuickTime™ and a GIF decompressor are needed to see this picture. How to improve CVA continued 3. Students: – Independently apply CVA. 172 QuickTime™ and a GIF decompressor are needed to see this picture. Developing Meaning Vocabulary in Classroom • Most important is building a culture that fosters interest in words. – Leads to word learning naturally. – Facilitates life-long interest and wordlearning abilities. 174 QuickTime™ and a GIF decompressor are needed to see this picture. Recommendations • Such classrooms have a culture that stimulates: – Word consciousness. – Word curiosity. – Word fun. – Wide reading. 175 QuickTime™ and a GIF decompressor are needed to see this picture. Rationales for Our Recommendation 1. We know learning words is natural! • Fully 90% (2,700±) of words learned per year are learned incidentally, outside of direct school instruction. • Directly teaching words results in a gain of few words–e.g., 50-100. 176 QuickTime™ and a GIF decompressor are needed to see this picture. Rationale for recommendations, continued • But facilitating attraction and attentiveness to words students hear or see in their normal daily life could increase natural word learning by 100s, if not 1000s. • These cognitive and affective factors promoted by such a culture will serve the student well for a lifetime. 177 QuickTime™ and a GIF decompressor are needed to see this picture. Rationale for recommendations, continued 2. We know a child’s environment profoundly affects vocabulary growth. 178 QuickTime™ and a GIF decompressor are needed to see this picture. Actual and Estimated Number of Words Heard from 0 - 48 Months 50 In Millions Estimated 40 Working Welfare 30 Actual 20 10 179 Age in Months 48 45 42 39 36 33 30 27 24 21 18 15 12 9 6 3 0 0 Actual & Estimated Words Addressed to Child Professional QuickTime™ and a GIF decompressor are needed to see this picture. Cumulative Number of New Words (Hart & Risley, 1995) 1200 1000 800 600 400 200 0 10 180 QuickTime™ and a GIF decompressor are needed to see this picture. QUESTIONS AND DISCUSSION 181 QuickTime™ and a GIF decompressor are needed to see this picture.