Beating Common Sense into Interactive Applications Henry Lieberman, Hugo Liu, Push Singh, Barbara Barry AI Magazine, Winter 2004 As (mis-)interpreted by Peter Clark for Boeing KR Group Introduction • Claim: Commonsense applications are closer than you think • Problems with CommonSense (CS) applications: – Even large KBs have sparse coverage – Inference is unreliable Their Common Sense KB: Open Mind Common Sense (OMCS) • 750k NL assertions from 15k contributors • ConceptNet: A semantic net built from these – 20 link types Against Question-Answering… • Question answering is a bad CS domain: – User expects a direct answer to all his/her questions – System has to be right (almost) all of the time – Got to be fast (few seconds) • Alternative: intelligent interfaces – Assists user when it can – “fail soft” - user can ignore it if he/she wants – But: Is yet another paperclip? 1. ARIA: Annotation and Retrieval Integration Agent • Helps annotate photos, and find photos – Similar to Thesaurus search – Photos are annotated with keywords – a. People, places and events are recognized in text – b. Use the semantic net to find “close” photos to text – Text also adds to the net (system learns) • “My sister’s name is Mary” → “Joe –sister→Mary” 2. Detecting Moods (“affect”) in Text “My wife left me; she took the kids and the dog” • Approach: – Mood keyword (e.g., “sad”) → mine a “small society of linguistic models of effect” from the KB (=?) • Applications: – Empathy Buddy: (purpose=?) – Summarizing a collection of reviews about a topic 3. Cinematic Commonsense: Video Capture and Editing • Videographer shoots, adds NL annotations – E.g., “a street artist is painting a painting” • Send annotations to KB for elaboration – “after painting, you clean the brushes” – “during painting, you might get paint on your hands” • Elaborations – suggest new shots for the videographer – Also are stored for improved retrieval – Can help order shots into temporal/causal • (isn’t temporal ordering already done?) • But: need more complex story understanding to create effective suggestions for the filmmaker. 4. Common Sense Storytelling: StoryIllustrator • Continuosly retreive photos relevant to user’s typing • Use Yahoo image search, not annotations, for Web images • CSK for query expansion, E.g., “baby” ↔ “milk” 5. Common Sense Storytelling: OMAdventure • Generates dungeons-and-dragons game on the fly • E.g., in kitchen → what do you find in kitchen? → Other associated locations? • E.g., oven → what can you do with an oven? • Hence oven, cooking are “moves” for user. Associated locations are “exits”. • User can add objects (e.g., “blender”) → extend KB (“blenders are in kitchens”) 7. Common Sense Storytelling: StoryFighter • System and user take turns to contribute lines to a story to get from A to B, e.g., – “John is sleepy” (start) – “John is in prison” (end) • Must avoid “taboo” words (e.g., “arrest”) • CSK deduces consequences of an event – “If you commit a crime, you might go to jail” • CSK also picks obvious taboo words 8. Topic Spotting • Task: Given speech, identify situation • E.g., “fries”, “lunch”, “Styrofoam” → “eating in a fast-food restaurant” • Use Bayesian inference + ConceptNet • Used in collaborative storytelling with kids – Computer starts the story – Kid continues – Computer can’t fully understand kid’s speech, but can at least identify the topic → generate plausible continuation • E.g., “bedroom” → “Jane’s parents walked into the bedroom while she was hiding under the bed” 9. Globuddy: A Tourist Phrasebook • Type in your situation – “I’ve just been arrested” • It retrieves and translates associated CS (?) – “If you are arrested, you should call a lawyer” – “Bail is a payment that allows an accused person to get out of jail until a trial” 10. Predictive typing/phrase completion • E.g., for a cellphone keyboard • Use ConceptNet to find next word that “makes sense” – E.g., “train st” → “train station” 11. Search: GOOSE and Reformulator as Google adjuncts • Infer user’s search goals and add keywords,e,g,: – “my cat is sick” → “Did you mean to look for veterinarians?” • Currently interactive. Later, will suggest better query. 12. Semantic Web • Given user’s goals, find services that might accomplish subgoals, e.g., – “Schedule a doctor’s appt” → look up directory of doctors, check reputation, geographic lookup, lookup schedules, etc. 13. Knowledge Acquisition • Criticism: Many OpenMind sentences are decontextualized – “At a wedding, bride and groom exchange rings” is culturally specific • → develop a prompt-based interface to have user’s make context explicit. Reflections • Logic: What inferences are possible • Commonsense: What inferences are plausible • Qn: How well does OpenMind support this? E.g., – “People live in houses” – “Things fall down rather than up” – “Acid irritates skin” Same for our own database… “There is a rocket” =? + Reflections (cont): Limitations • Spottiness of subject coverage in OpenMind • Inference is unreliable → reluctant to use it – Need new inference methods – E.g., “interleave context-sensitive inference with retrieval in a breadth-first manner” • CS suggestions may be distracting – But trials suggest otherwise (people tolerate wrong but plausible suggestions better than stupid ones) Some additional thoughts… • Domain-specific vs. domain-general applications – Domain-specific – how much CS is needed? • CycSecure • Oil exploration • etc. – Domain-general – still need task-specific algorithm – Unusual to find a domain- and task-general application • “Scenario completion” is a good task – newswire, incident reports, etc.