Is Question Answering an Acquired Skill? Ramakrishnan, Chakrabarti, Paranjpe, Bhattacharyya Paper presentation: Vinay Goel Introduction • Question Answering (QA) system • Most QA systems are substantial team efforts – Difficult to reproduce a well tuned QA system from scratch, gauge the benefit of new algorithmic ideas, new corpora and new languages • QA systems – Complex building blocks like taggers and parsers – Lashed together with customized “glue” – Crucial knobs best preset by QA specialists Goal • Decompose the QA task cleanly into discovering features and learning to score answer snippets • QA system – Performs fast, shallow processing of corpus – Structures the scoring task using features and learners – Trains its scoring algorithm from a past history of questions and vetted answers – Can include side information (Wordnet etc.) – Reuses expertise accumulated from one corpus to a new corpus Noisy simulation perspective • In a structured database with a suitable schema and a structured query language, information needs can be expressed clearly • See QA as a transformation of this process by adding natural language from an unknown generative process, for both query and data • Given a question, discover structured fragments in it – Extract selectors which will appear almost unchanged in an answer passage – Extract atype clues, which tell what else to look for in a passage that satisfied all selectors Typical connections between a question and answer Atype • Minimal subclass of entities which will answer a question • Two representations important to factoid QA – Atype as synset – Atype as surface patterns Atype as synset • Q: Name an animal that sleeps upright • A: horse • Wordnet helps recognize that horse is an instance of animal • Most answers which are common nouns are assisted by this representation Atype as surface patterns • Infinite or very large domains such as numbers, person names, place names etc. cannot be covered by Wordnet • Logically augment Wordnet to add connections from synsets to pattern matchers such as “at DD:DD” or “Xx+ said” etc. From the question to an atype • Set of common “wh-words” • Questions starting with when, where and who immediately reveal their expected atypes • Word after how is almost always a clue • Questions using using what and which mention atype directly Shallow parsing to extract atype • Shallow parsing involves finding noun phrases, modifiers and attachments between phrases • Purely based on POS tags • Strategy for locating atype clues from “what” and “which” questions: – Head of NP appearing before the auxiliary / main verb if it is not a wh-word – Otherwise, head of NP appearing after the verb Learning to map atype • When, where, who and how do not directly use a term that describes a synset • Augmented synsets based on surface patterns (DDDD) may come handy • Devised a learning module to help compile mappings between short token sequences in questions to atypes Selectors • Second blank in the SQL query select…where… • In QA, simply a set of words in the question that are expected to appear unchanged in the answer passage Identifying selectors • Choice of features – POS – POS assigned to left and right neighbors – Whether the word starts with an uppercase letter – – – – Whether the word is a stopword Some version of IDF How many senses the word has in isolation For a given sense, how many other words describe the sense How to use selectors • Two places – Pad the initial keyword query – Rerank the candidate phases • Experiments insist that the response by the word search engine (Lucene) – Contains all selectors – Use OR over other question words Learning to score passages • If (q,r) is a positive instance, it is expected that – All selectors match between q and r – r has an answer zone a which does not contain selectors – The linear distance between a and matched selectors in r, tend to be small – a has strong Wordnet-based similarity with the atype of q Overall architecture Experiments • TREC QA track • Picked sliding windows of three sentences as passages • Questions and passages were tokenized using GATE • For learning tasks, used J48 decision tree and the logistic regression packages in WEKA Extracting atypes from shallow parses Spotting selectors in questions Passage Reranking performance MRR improvement via reranking Training on corpus of another year Conclusion • QA system built by wrapping logic around text indexers, taggers, shallow parsers and classifiers • Simple assembly of building blocks • Future work involves improving performance of different blocks