A Semantic Parsing and Reasoning-Based Approach to Knowledge Base Question Answering

Total Page:16

File Type:pdf, Size:1020Kb

A Semantic Parsing and Reasoning-Based Approach to Knowledge Base Question Answering The Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI-21) A Semantic Parsing and Reasoning-Based Approach to Knowledge Base Question Answering Ibrahim Abdelaziz∗, Srinivas Ravishankar∗, Pavan Kapanipathi∗, Salim Roukos, Alexander Gray IBM Research IBM T.J. Watson Research Center Yorktown Heights, NY, USA fibrahim.abdelaziz1, srini, [email protected], fkapanipa, [email protected] Abstract associated PropBank information to a KB such as DBpedia. Finally, the logic forms are used by Logical Neural Network Knowledge Base Question Answering (KBQA) is a task where existing techniques have faced significant challenges, such as (LNN) (Riegel et al. 2020), a neuro-symbolic reasoner that the need for complex question understanding, reasoning, and enables DTQA to perform different types of reasoning. DTQA large training datasets. In this work, we demonstrate Deep comprises several modules which operate within a pipeline Thinking Question Answering (DTQA), a semantic parsing and are individually trained for their specific tasks. and reasoning-based KBQA system. DTQA (1) integrates mul- This demonstration shows an interpretable methodology tiple, reusable modules that are trained specifically for their for KBQA that achieves state-of-the-art results on two promi- individual tasks (e.g. semantic parsing, entity linking, and rela- nent datasets (QALD-9, LC-QuAD-1). DTQA is a system of tionship linking), eliminating the need for end-to-end KBQA systems that demonstrates: (1) the use of task-general seman- training data; (2) leverages semantic parsing and a reasoner for tic parsing via AMR; (2) an approach that aligns AMR to KB improved question understanding. DTQA is a system of sys- triples; and (3) the use of a neuro-symbolic reasoner for the tems that achieves state-of-the-art performance on two popular KBQA datasets. KBQA task. DTQA System Introduction Figure 1 shows an overview of DTQA with an example ques- The goal of Knowledge Base Question Answering is to an- tion. The input to DTQA is the question text, which is first swer natural language questions over Knowledge Bases (KB) parsed using the Abstract Meaning Representation (AMR) such as DBpedia (Auer et al. 2007) and Wikidata (Vrandeciˇ c´ parser (Banarescu et al. 2013; Dorr, Habash, and Traum and Krotzsch¨ 2014). Existing approaches to KBQA are ei- 1998). AMR is a semantic parse representation that solves ther (i) graph-driven (Diefenbach et al. 2020; Vakulenko the ambiguity of natural language by representing syntacti- et al. 2019), or (ii) end-to-end learning (Sun et al. 2020) cally different sentences with the same underlying meaning approaches. Graph-driven approaches attempt to find a best- in the same way. An AMR parse of a sentence is a rooted, match KB sub-graph for the given question, whereas end-to- directed, acyclic graph expressing “who is doing what to end learning approaches, requiring large amounts of training whom”. We use the current state-of-the-art system for AMR data, directly predict a query sequence (i.e. SPARQL or SQL) that leverage transition-based (Naseem et al. 2019) parsing from the input question. Both categories still suffer from a approach, parameterized with neural networks and trained number of challenges, including complex question under- in an end-to-end fashion. AMR parses are KB independent, standing, scarcity of end-to-end training data, and the need however an essential task to solve KBQA using AMR is to for reasoning. align the AMR parses with the knowledge base. Therefore, in In this work, we demonstrate DTQA; a semantic-parsing the next three modules we align the AMR parse to a DBpedia and reasoning-based KBQA system. DTQA begins by parsing sub-graph that can be transformed into a SPARQL query. the natural language question into Abstract Meaning Repre- First, DTQA uses an Entity Extraction and Linking (EL) sentation (AMR), which is a symbolic formalism that cap- module that extracts entities and types and links them to their tures rich semantic information from natural language. The corresponding KB URIs. Given a question such as Which ac- use of AMR enables task-independent language understand- tors starred in Spanish movies produced by Benicio del Toro?, ing, which DTQA leverages to answer structurally complex the EL module will link the AMR nodes of Benicio del Toro, questions. Next, our system uses a triples-based approach and Spanish to dbr:Benicio del Toro and dbr:Spain, respec- that builds on state-of-the-art entity and relation linking mod- tively. In DTQA, we trained a BERT-based neural mention ules to align the AMR graph with the underlying KB and detection and used BLINK (Devlin et al. 2018) for disam- produce logic forms. This step accurately maps AMR and its biguation of named entities. ∗Equal contribution Aligning the AMR graph to the underlying KB is a chal- Copyright c 2021, Association for the Advancement of Artificial lenge, particularly due to structural and granularity mismatch Intelligence (www.aaai.org). All rights reserved. between AMR and KBs such as DBPedia. The PropBank 15985 !"#$%&'()*$+,'%- ."$,/%'%- 4+56,$ !"#$,3$ /31-8&2$9)6'&2$ !"#$%&'()' *+,-,.$/-+0-+1 !"#$,3$=>$?'-52)( #)2&,-3+$/-+0-+1 76)(,-3+ /31-8 9),:3'0($ ;/99< #,$ "-./'0' &()*+' #1"%&' !"#$%&' +,-"()* arg ! (type(x, dbo:Film) ^ country(x, dbr:Spain) ^ -"2) -"2* 0+:+*;+ ,*-.) z -"23 !"#$*+,&&-./ producer(x, dbr:Benicio_del_Toro) ^ ()$!*+ ,*-.) starring(x, z) ^ type(z, Person) ) .#$ -"2* 0+:+*;+ .#$ .#$/&#%0,"1 ($*/01)%&' 1*0+"$8 9(#.+ !"#$(#'.+&0 -"2) 0-.' "-./'1' !"#$%&' 29(#.+7 +,-"()* SELECT DISTINCT ?uri WHERE { ()$!*+ 0+:+*;+ ,*-.) Which actors 0-.' !"#$*+,&&-./ ?film rdf:type dbo:Film ; dbo:country dbr:Spain ; starred in Spanish 23)+.1.* !"#$./0(' ()* dbo:producer dbr:Benicio_del_Toro ; movies produced /)456*$*7 dbo:starring ?uri . !"#$%&'%&#!'()& by Benicio del 3)+.1.* $ } Toro? !"#$%&'()(*+!&,+-*#* /)456*$* !"# Figure 1: DTQA architecture, illustrated via examples frames in AMR are n-ary predicates whereas relationships mapping of an AMR predicate to its semantically equivalent in KBs are predominantly binary. In order to transform the relation from the underlying KB using a distant supervision n-ary predicates to binary, we use a novel ‘Path-based Triple dataset. Apart from this statistical co-occurrence informa- Generation’ module that transforms the AMR graph (with tion, the rich lexical information in the AMR predicate and linked entities) into a set of triples where each has one to question text is also leveraged by SLING. This is captured one correspondence with constraints in the SPARQL query. by a neural-model and an unsupervised semantic similarity The main idea of this module is to find paths between the module that measures similarity of each AMR predicate with question amr-unknown to every linked entity and hence avoid a list of possible relations using GloVe (Pennington, Socher, generating unnecessary triples. For the example in Figure 1, and Manning 2014) embeddings. this approach will generate the following two paths: The AMR to Logic module is a rule-based system that 1. amr-unknown ! star-01 ! movie ! country ! Spain transforms the KB-aligned AMR paths to a logical formalism 2. amr-unknown ! star-01 ! movie ! produce-01 ! Beni- to support binary predicates and higher-order functional pred- cio del Toro. icates (e.g. for aggregation and manipulation of sets). This logic formalism has a deterministic mapping to SPARQL by Note that the predicate act-01 in the AMR graph is irrel- introducing SPARQL constructs such as COUNT, ORDER evant, since it is not on the path between amr-unknown and BY, and FILTER. the entities, and hence ignored. Some questions in KBQA datasets require reasoning. How- Furthermore, AMR parse provides relationship informa- ever, existing KBQA systems (Diefenbach et al. 2020; Vaku- tion that is generally more fine-grained than the KB. To lenko et al. 2019; Sun et al. 2020) do not have a modular address this granularity mismatch, the path-based triple gen- reasoner to perform logical reasoning using the information eration collapses multiple predicates occurring consecutively available in the knowledge base. In DTQA, the final mod- into a single edge. For example, in the question “In which ule is a logical reasoner called Logical Neural Networks ancient empire could you pay with cocoa beans? “, the path (LNN) (Riegel et al. 2020). LNN is a neural network architec- consists of three predicates between amr-unknown and cocoa ture in which neurons model a rigorously defined notion of bean; (location, pay-01, instrument). These are collapsed into weighted fuzzy or classical first-order logic. LNN takes as in- a single edge, and the result is amr-unknown ! location j put the logic forms from AMR to Logic and has access to the pay-01j instrument ! cocoa-bean. KB to provide an answer. LNN retrieves predicate groundings Next, the (collapsed) AMR predicates from the triples are via its granular SPARQL integration and performs type-based, mapped to relationships in the KB using the Relation Extrac- and geographic reasoning based on the ontological informa- tion and Linking (REL) module. As seen in Figure 1, the tion available in the knowledge base. LNN allowed us to REL module maps star-01 and produce-01 to DBpedia rela- manually add generic axioms for the geographic reasoning tions dbo:starring and dbo:producer, respectively. For this scenarios. task, DTQA uses SLING (Mihindukulasooriya et al. 2020); the state-of-the-art REL system for KBQA. SLING takes as input the AMR graph with linked entities and extracted paths Experimental Evaluation obtained from the upstream modules. It relies on an ensem- To evaluate DTQA, we use two prominent KBQA datasets; ble of complementary relation linking approaches to produce QALD - 9 (Ngomo 2018) and LC-QuAD 1.0 (Trivedi et al. a top-k list of KB relations for each AMR predicate. The 2017). The latest version of QALD-9 contains 150 test ques- most relevant and novel approach in SLING is its statistical tions whereas LC-QuAD test set contains 1,000 questions.
Recommended publications
  • Abstraction and Generalisation in Semantic Role Labels: Propbank
    Abstraction and Generalisation in Semantic Role Labels: PropBank, VerbNet or both? Paola Merlo Lonneke Van Der Plas Linguistics Department Linguistics Department University of Geneva University of Geneva 5 Rue de Candolle, 1204 Geneva 5 Rue de Candolle, 1204 Geneva Switzerland Switzerland [email protected] [email protected] Abstract The role of theories of semantic role lists is to obtain a set of semantic roles that can apply to Semantic role labels are the representa- any argument of any verb, to provide an unam- tion of the grammatically relevant aspects biguous identifier of the grammatical roles of the of a sentence meaning. Capturing the participants in the event described by the sentence nature and the number of semantic roles (Dowty, 1991). Starting from the first proposals in a sentence is therefore fundamental to (Gruber, 1965; Fillmore, 1968; Jackendoff, 1972), correctly describing the interface between several approaches have been put forth, ranging grammar and meaning. In this paper, we from a combination of very few roles to lists of compare two annotation schemes, Prop- very fine-grained specificity. (See Levin and Rap- Bank and VerbNet, in a task-independent, paport Hovav (2005) for an exhaustive review). general way, analysing how well they fare In NLP, several proposals have been put forth in in capturing the linguistic generalisations recent years and adopted in the annotation of large that are known to hold for semantic role samples of text (Baker et al., 1998; Palmer et al., labels, and consequently how well they 2005; Kipper, 2005; Loper et al., 2007). The an- grammaticalise aspects of meaning.
    [Show full text]
  • A Novel Large-Scale Verbal Semantic Resource and Its Application to Semantic Role Labeling
    VerbAtlas: a Novel Large-Scale Verbal Semantic Resource and Its Application to Semantic Role Labeling Andrea Di Fabio}~, Simone Conia}, Roberto Navigli} }Department of Computer Science ~Department of Literature and Modern Cultures Sapienza University of Rome, Italy {difabio,conia,navigli}@di.uniroma1.it Abstract The automatic identification and labeling of ar- gument structures is a task pioneered by Gildea We present VerbAtlas, a new, hand-crafted and Jurafsky(2002) called Semantic Role Label- lexical-semantic resource whose goal is to ing (SRL). SRL has become very popular thanks bring together all verbal synsets from Word- to its integration into other related NLP tasks Net into semantically-coherent frames. The such as machine translation (Liu and Gildea, frames define a common, prototypical argu- ment structure while at the same time pro- 2010), visual semantic role labeling (Silberer and viding new concept-specific information. In Pinkal, 2018) and information extraction (Bas- contrast to PropBank, which defines enumer- tianelli et al., 2013). ative semantic roles, VerbAtlas comes with In order to be performed, SRL requires the fol- an explicit, cross-frame set of semantic roles lowing core elements: 1) a verb inventory, and 2) linked to selectional preferences expressed in terms of WordNet synsets, and is the first a semantic role inventory. However, the current resource enriched with semantic information verb inventories used for this task, such as Prop- about implicit, shadow, and default arguments. Bank (Palmer et al., 2005) and FrameNet (Baker We demonstrate the effectiveness of VerbAtlas et al., 1998), are language-specific and lack high- in the task of dependency-based Semantic quality interoperability with existing knowledge Role Labeling and show how its integration bases.
    [Show full text]
  • Handling Ambiguity Problems of Natural Language Interface For
    IJCSI International Journal of Computer Science Issues, Vol. 9, Issue 3, No 3, May 2012 ISSN (Online): 1694-0814 www.IJCSI.org 17 Handling Ambiguity Problems of Natural Language InterfaceInterface for QuesQuesQuestQuestttionionionion Answering Omar Al-Harbi1, Shaidah Jusoh2, Norita Norwawi3 1 Faculty of Science and Technology, Islamic Science University of Malaysia, Malaysia 2 Faculty of Science & Information Technology, Zarqa University, Zarqa, Jordan 3 Faculty of Science and Technology, Islamic Science University of Malaysia, Malaysia documents processing, and answer processing [9], [10], Abstract and [11]. The Natural language question (NLQ) processing module is considered a fundamental component in the natural language In QA, a NLQ is the primary source through which a interface of a Question Answering (QA) system, and its quality search process is directed for answers. Therefore, an impacts the performance of the overall QA system. The most accurate analysis to the NLQ is required. One of the most difficult problem in developing a QA system is so hard to find an exact answer to the NLQ. One of the most challenging difficult problems in developing a QA system is that it is problems in returning answers is how to resolve lexical so hard to find an answer to a NLQ [22]. The main semantic ambiguity in the NLQs. Lexical semantic ambiguity reason is most QA systems ignore the semantic issue in may occurs when a user's NLQ contains words that have more the NLQ analysis [12], [2], [14], and [15]. To achieve a than one meaning. As a result, QA system performance can be better performance, the semantic information contained negatively affected by these ambiguous words.
    [Show full text]
  • The Question Answering System Using NLP and AI
    International Journal of Scientific & Engineering Research Volume 7, Issue 12, December-2016 ISSN 2229-5518 55 The Question Answering System Using NLP and AI Shivani Singh Nishtha Das Rachel Michael Dr. Poonam Tanwar Student, SCS, Student, SCS, Student, SCS, Associate Prof. , SCS Lingaya’s Lingaya’s Lingaya’s Lingaya’s University, University,Faridabad University,Faridabad University,Faridabad Faridabad Abstract: (ii) QA response with specific answer to a specific question The Paper aims at an intelligent learning system that will take instead of a list of documents. a text file as an input and gain knowledge from the given text. Thus using this knowledge our system will try to answer questions queried to it by the user. The main goal of the 1.1. APPROCHES in QA Question Answering system (QAS) is to encourage research into systems that return answers because ample number of There are three major approaches to Question Answering users prefer direct answers, and bring benefits of large-scale Systems: Linguistic Approach, Statistical Approach and evaluation to QA task. Pattern Matching Approach. Keywords: A. Linguistic Approach Question Answering System (QAS), Artificial Intelligence This approach understands natural language texts, linguistic (AI), Natural Language Processing (NLP) techniques such as tokenization, POS tagging and parsing.[1] These are applied to reconstruct questions into a correct 1. INTRODUCTION query that extracts the relevant answers from a structured IJSERdatabase. The questions handled by this approach are of Factoid type and have a deep semantic understanding. Question Answering (QA) is a research area that combines research from different fields, with a common subject, which B.
    [Show full text]
  • Quester: a Speech-Based Question Answering Support System for Oral Presentations Reza Asadi, Ha Trinh, Harriet J
    Quester: A Speech-Based Question Answering Support System for Oral Presentations Reza Asadi, Ha Trinh, Harriet J. Fell, Timothy W. Bickmore Northeastern University Boston, USA asadi, hatrinh, fell, [email protected] ABSTRACT good support for speakers who want to deliver more Current slideware, such as PowerPoint, reinforces the extemporaneous talks in which they dynamically adapt their delivery of linear oral presentations. In settings such as presentation to input or questions from the audience, question answering sessions or review lectures, more evolving audience needs, or other contextual factors such as extemporaneous and dynamic presentations are required. varying or indeterminate presentation time, real-time An intelligent system that can automatically identify and information, or more improvisational or experimental display the slides most related to the presenter’s speech, formats. At best, current slideware only provides simple allows for more speaker flexibility in sequencing their indexing mechanisms to let speakers hunt through their presentation. We present Quester, a system that enables fast slides for material to support their dynamically evolving access to relevant presentation content during a question speech, and speakers must perform this frantic search while answering session and supports nonlinear presentations led the audience is watching and waiting. by the speaker. Given the slides’ contents and notes, the system ranks presentation slides based on semantic Perhaps the most common situations in which speakers closeness to spoken utterances, displays the most related must provide such dynamic presentations are in Question slides, and highlights the corresponding content keywords and Answer (QA) sessions at the end of their prepared in slide notes. The design of our system was informed by presentations.
    [Show full text]
  • Generating Lexical Representations of Frames Using Lexical Substitution
    Generating Lexical Representations of Frames using Lexical Substitution Saba Anwar Artem Shelmanov Universitat¨ Hamburg Skolkovo Institute of Science and Technology Germany Russia [email protected] [email protected] Alexander Panchenko Chris Biemann Skolkovo Institute of Science and Technology Universitat¨ Hamburg Russia Germany [email protected] [email protected] Abstract Seed sentence: I hope PattiHelper can helpAssistance youBenefited party soonTime . Semantic frames are formal linguistic struc- Substitutes for Assistance: assist, aid tures describing situations/actions/events, e.g. Substitutes for Helper: she, I, he, you, we, someone, Commercial transfer of goods. Each frame they, it, lori, hannah, paul, sarah, melanie, pam, riley Substitutes for Benefited party: me, him, folk, her, provides a set of roles corresponding to the sit- everyone, people uation participants, e.g. Buyer and Goods, and Substitutes for Time: tomorrow, now, shortly, sooner, lexical units (LUs) – words and phrases that tonight, today, later can evoke this particular frame in texts, e.g. Sell. The scarcity of annotated resources hin- Table 1: An example of the induced lexical represen- ders wider adoption of frame semantics across tation (roles and LUs) of the Assistance FrameNet languages and domains. We investigate a sim- frame using lexical substitutes from a single seed sen- ple yet effective method, lexical substitution tence. with word representation models, to automat- ically expand a small set of frame-annotated annotated resources. Some publicly available re- sentences with new words for their respective sources are FrameNet (Baker et al., 1998) and roles and LUs. We evaluate the expansion PropBank (Palmer et al., 2005), yet for many lan- quality using FrameNet.
    [Show full text]
  • Ontology-Based Approach to Semantically Enhanced Question Answering for Closed Domain: a Review
    information Review Ontology-Based Approach to Semantically Enhanced Question Answering for Closed Domain: A Review Ammar Arbaaeen 1,∗ and Asadullah Shah 2 1 Department of Computer Science, Faculty of Information and Communication Technology, International Islamic University Malaysia, Kuala Lumpur 53100, Malaysia 2 Faculty of Information and Communication Technology, International Islamic University Malaysia, Kuala Lumpur 53100, Malaysia; [email protected] * Correspondence: [email protected] Abstract: For many users of natural language processing (NLP), it can be challenging to obtain concise, accurate and precise answers to a question. Systems such as question answering (QA) enable users to ask questions and receive feedback in the form of quick answers to questions posed in natural language, rather than in the form of lists of documents delivered by search engines. This task is challenging and involves complex semantic annotation and knowledge representation. This study reviews the literature detailing ontology-based methods that semantically enhance QA for a closed domain, by presenting a literature review of the relevant studies published between 2000 and 2020. The review reports that 83 of the 124 papers considered acknowledge the QA approach, and recommend its development and evaluation using different methods. These methods are evaluated according to accuracy, precision, and recall. An ontological approach to semantically enhancing QA is found to be adopted in a limited way, as many of the studies reviewed concentrated instead on Citation: Arbaaeen, A.; Shah, A. NLP and information retrieval (IR) processing. While the majority of the studies reviewed focus on Ontology-Based Approach to open domains, this study investigates the closed domain.
    [Show full text]
  • A Question-Driven News Chatbot
    What’s The Latest? A Question-driven News Chatbot Philippe Laban John Canny Marti A. Hearst UC Berkeley UC Berkeley UC Berkeley [email protected] [email protected] [email protected] Abstract a set of essential questions and link each question with content that answers it. The motivating idea This work describes an automatic news chat- is: two pieces of content are redundant if they an- bot that draws content from a diverse set of news articles and creates conversations with swer the same questions. As the user reads content, a user about the news. Key components of the system tracks which questions are answered the system include the automatic organization (directly or indirectly) with the content read so far, of news articles into topical chatrooms, inte- and which remain unanswered. We evaluate the gration of automatically generated questions system through a usability study. into the conversation, and a novel method for The remainder of this paper is structured as fol- choosing which questions to present which lows. Section2 describes the system and the con- avoids repetitive suggestions. We describe the algorithmic framework and present the results tent sources, Section3 describes the algorithm for of a usability study that shows that news read- keeping track of the conversation state, Section4 ers using the system successfully engage in provides the results of a usability study evaluation multi-turn conversations about specific news and Section5 presents relevant prior work. stories. The system is publicly available at https:// newslens.berkeley.edu/ and a demonstration 1 Introduction video is available at this link: https://www.
    [Show full text]
  • Natural Language Processing with Deep Learning Lecture Notes: Part Vii Question Answering 2 General QA Tasks
    CS224n: Natural Language Processing with Deep 1 Learning 1 Course Instructors: Christopher Lecture Notes: Part VII Manning, Richard Socher 2 Question Answering 2 Authors: Francois Chaubard, Richard Socher Winter 2019 1 Dynamic Memory Networks for Question Answering over Text and Images The idea of a QA system is to extract information (sometimes pas- sages, or spans of words) directly from documents, conversations, online searches, etc., that will meet user’s information needs. Rather than make the user read through an entire document, QA system prefers to give a short and concise answer. Nowadays, a QA system can combine very easily with other NLP systems like chatbots, and some QA systems even go beyond the search of text documents and can extract information from a collection of pictures. There are many types of questions, and the simplest of them is factoid question answering. It contains questions that look like “The symbol for mercuric oxide is?” “Which NFL team represented the AFC at Super Bowl 50?”. There are of course other types such as mathematical questions (“2+3=?”), logical questions that require extensive reasoning (and no background information). However, we can argue that the information-seeking factoid questions are the most common questions in people’s daily life. In fact, most of the NLP problems can be considered as a question- answering problem, the paradigm is simple: we issue a query, and the machine provides a response. By reading through a document, or a set of instructions, an intelligent system should be able to answer a wide variety of questions. We can ask the POS tags of a sentence, we can ask the system to respond in a different language.
    [Show full text]
  • Distributional Semantics, Wordnet, Semantic Role
    Semantics Kevin Duh Intro to NLP, Fall 2019 Outline • Challenges • Distributional Semantics • Word Sense • Semantic Role Labeling The Challenge of Designing Semantic Representations • Q: What is semantics? • A: The study of meaning • Q: What is meaning? • A: … We know it when we see it • These sentences/phrases all have the same meaning: • XYZ corporation bought the stock. • The stock was bought by XYZ corporation. • The purchase of the stock by XYZ corporation... • The stock purchase by XYZ corporation... But how to formally define it? Meaning Sentence Representation Example Representations Sentence: “I have a car” Logic Formula Graph Representation Key-Value Records Example Representations Sentence: “I have a car” As translation in “Ich habe ein Auto” another language There’s no single agreed-upon representation that works in all cases • Different emphases: • Words or Sentences • Syntax-Semantics interface, Logical Inference, etc. • Different aims: • “Deep (and narrow)” vs “Shallow (and broad)” • e.g. Show me all flights from BWI to NRT. • Do we link to actual flight records? • Or general concept of flying machines? Outline • Challenges • Distributional Semantics • Word Sense • Semantic Role Labeling Distributional Semantics 10 Learning Distributional Semantics from large text dataset • Your pet dog is so cute • Your pet cat is so cute • The dog ate my homework • The cat ate my homework neighbor(dog) overlaps-with neighbor(cats) so meaning(dog) is-similar-to meaning(cats) 11 Word2Vec implements Distribution Semantics Your pet dog is so cute 1. Your pet - is so | dog 2. dog | Your pet - is so From: Mikolov, Tomas; et al. (2013). "Efficient Estimation of Word 12 Representations in Vector Space” Latent Semantic Analysis (LSA) also implements Distributional Semantics Pet-peeve: fundamentally, neural approaches aren’t so different from classical LSA.
    [Show full text]
  • Structured Retrieval for Question Answering
    SIGIR 2007 Proceedings Session 14: Question Answering Structured Retrieval for Question Answering Matthew W. Bilotti, Paul Ogilvie, Jamie Callan and Eric Nyberg Language Technologies Institute Carnegie Mellon University 5000 Forbes Avenue Pittsburgh, PA 15213 { mbilotti, pto, callan, ehn }@cs.cmu.edu ABSTRACT keywords in a minimal window (e.g., Ruby killed Oswald is Bag-of-words retrieval is popular among Question Answer- not an answer to the question Who did Oswald kill?). ing (QA) system developers, but it does not support con- Bag-of-words retrieval does not support checking such re- straint checking and ranking on the linguistic and semantic lational constraints at retrieval time, so many QA systems information of interest to the QA system. We present an simply query using the key terms from the question, on the approach to retrieval for QA, applying structured retrieval assumption that any document relevant to the question must techniques to the types of text annotations that QA systems necessarily contain them. Slightly more advanced systems use. We demonstrate that the structured approach can re- also index and provide query operators that match named trieve more relevant results, more highly ranked, compared entities, such as Person and Organization [14]. When the with bag-of-words, on a sentence retrieval task. We also corpus includes many documents containing the keywords, characterize the extent to which structured retrieval effec- however, the ranked list of results can contain many irrel- tiveness depends on the quality of the annotations. evant documents. For example, question 1496 from TREC 2002 asks, What country is Berlin in? A typical query for this question might be just the keyword Berlin,orthekey- Categories and Subject Descriptors word Berlin in proximity to a Country named entity.
    [Show full text]
  • Foundations of Natural Language Processing Lecture 16 Semantic
    Language is Flexible Foundations of Natural Language Processing Often we want to know who did what to whom (when, where, how and why) • Lecture 16 But the same event and its participants can have different syntactic realizations. Semantic Role Labelling and Argument Structure • Sandy broke the glass. vs. The glass was broken by Sandy. Alex Lascarides She gave the boy a book. vs. She gave a book to the boy. (Slides based on those of Schneider, Koehn, Lascarides) Instead of focusing on syntax, consider the semantic roles (also called 13 March 2020 • thematic roles) defined by each event. Alex Lascarides FNLP Lecture 16 13 March 2020 Alex Lascarides FNLP Lecture 16 1 Argument Structure and Alternations Stanford Dependencies Mary opened the door • The door opened John slices bread with a knife • This bread slices easily The knife slices cleanly Mary loaded the truck with hay • Mary loaded hay onto the the truck The truck was loaded with hay (by Mary) The hay was loaded onto the truck (by Mary) John gave a present to Mary • John gave Mary a present cf Mary ate the sandwich with Kim! Alex Lascarides FNLP Lecture 16 2 Alex Lascarides FNLP Lecture 16 3 Syntax-Semantics Relationship Outline syntax = semantics • 6 The semantic roles played by different participants in the sentence are not • trivially inferable from syntactical relations . though there are patterns! • The idea of semantic roles can be combined with other aspects of meaning • (beyond this course). Alex Lascarides FNLP Lecture 16 4 Alex Lascarides FNLP Lecture 16 5 Commonly used thematic
    [Show full text]