CS474 Intro to Natural Language Question answering Processing • Overview and task definition • History • Open-domain question answering Question Answering • Basic system architecture • Predictive indexing methods – Slides based on those of Jamie Callan, CMU • Pattern-matching methods • Advanced techniques
Indexing with predictive annotation Predictive annotation • Some answers belong to well-defined semantic classes – People, places, monetary amounts, telephone numbers, addresses, organizations • Predictive annotation: index a document with “concepts” or “features” that are expected to be useful in (many) queries – E.g. people names, location names, addresses, etc. • Add additional operators for use in queries – E.g. Where does Ellen Vorhees work? “Ellen Vorhees” NEAR/10 *organization Predictive annotation Issues for predictive annotation
• How is annotated text stored in the index? • What makes a good QA-token? – Question that would use the token • Can be recognized with high reliability (high precision) • Occurs frequently enough to be worth the effort • How do you want the system to make use of the QA-tokens? – Filtering step? – Transform original question into an ad-hoc retrieval question that incorporates QA-tokens and proximity operators? • Common approaches to recognizing QA-tokens • Treat <$QA-token, term> as meaning that $QA-token – Tables, lists, dictionaries and term occur at the same location in the text – Heuristics – Or use phrase indexing approach to index as a single – Hidden Markov models item
Advantages and disadvantages Question answering + Most of the computational cost occurs during indexing • Overview and task definition • Allows use of more sophisticated methods • History + Annotator has access to complete text of document • Important for recognizing some types of features • Open-domain question answering – Must know ahead of time which types of • Basic system architecture features/concepts are likely to be important – Increases size of index considerably • Predictive indexing methods • E.g. by an order of magnitude if many features • Pattern-matching methods – Slides based on those of Jamie Callan, CMU • Used (in varying amounts) by almost all open-domain • Advanced techniques Q/A systems Simple pattern-based QA Example • Observation: there are many • Who is the President of Cornell? questions…but fewer types of questions – Expectation: answer string contains a • Each type of question can be associated person name with • Named entity identification – Expectations about answer string – Search query: “president Cornell characteristics *PersonName” – Strategies for retrieving documents that – Rule: “*PersonName, President of Cornell” might have answers • Matches “…Hunter Rawlings, President of –Rulesfor identifying answer strings in Cornell” documents • Answer = “Hunter Rawlings”
Question analysis Question analysis example • Input: the question • “Who is Elvis?” • Output – Question type: “who” – Named-entity tagging: “Who is
Simple pattern-based Q/A: Answer extraction assessment • Select highly ranked sentences from highly ranked documents • Extremely effective when • Perform named-entity tagging (or extract from index) – Question patterns are predictable and perform part of speech tagging • Fairly “few” patterns cover the most likely questions – Could be several hundred – “The/DT planet/NN
Common problem: improving the Common problem: improving set of retrieved documents answer extraction patterns
• Sometimes the IR system can’t find any documents that • Word sequence patterns have limited power have answers (even though the right documents are in the corpus) • Solution: create patterns that use syntactic information • Solution: get a broader set of documents – Partial syntactic parsing of documents – Approach: if answer extractor fails to find an answer, • Is this noun the subject or the object of the sentence? kick the question back to the search engine with – Allows more complex patterns instructions to widen the search • Question: “Who shot Kennedy?” • Assumes answer extractors can tell when they fail • “Who” implies a person that should be subject of answer – Approach: use a variety of retrieval strategies to retrieve sentence/clause documents • “Kennedy” should be direct object of answer • E.g., all words within one sentence, then all words within one • Pattern:
SMU/LCC system Multi-strategy approach • State of the art in QA is the SMU/LCC Falcon system – Employs informed use of standard IR techniques – Use of broad ontology (extended WordNet) TREC 2001 – Lots of NLP – Answer verification • Similar to most other systems in architecture TREC 2000 except for – Much more careful tuning of algorithms and resources – More sophisticated control of IR and NLP – Feedback loops Question analysis Expected answer types
• Parsing and named entity recognition • Expected answer type determined by parsing
• Exceptions for “special cases”
Expected answer types Paragraph retrieval • Answer types are mapped to named-entity • Boolean retrieval with loops categories that can be recognized in text – Different from multiple queries in that system only uses additional queries when necessary – Fewer candidates for analysis components to consider • Loop 1: query keyword loop – Keywords added/dropped to make query more/less specific • Loop 2: keyword alternations – Try morphological variants and synonyms • Answer types drive processing of paragraphs • Loop 3: semantic alternations – Passages need to contain the expected answer type – Try semantic alternatives Feedback loops Answer verification • Parse passages to create a dependency tree among words • Attempt to unify logical forms of question and answer text
Assessment
• Strengths – Controlled use of IR system • Query expansion via lexical and semantic equivalents • Believed to be the major power of the system – Tailored resources (see paper) • WordNet, parser, NE identifier, etc. – Answer verification • Initially thought to be the key component of the system •Now…not so clear • Weaknesses – Complex system, contribution of each component unclear