Paraphrase and Textual Entailment Recognition And

Total Page:16

File Type:pdf, Size:1020Kb

Paraphrase and Textual Entailment Recognition And PARAPHRASE AND TEXTUAL ENTAILMENT RECOGNITION AND GENERATION Prodromos Malakasiotis PH.D. THESIS DEPARTMENT OF INFORMATICS ATHENS UNIVERSITY OF ECONOMICS AND BUSINESS 2011 Abstract Paraphrasing methods recognize, generate, or extract phrases, sentences, or longer nat- ural language expressions that convey almost the same information. Textual entailment methods, on the other hand, recognize, generate, or extract pairs of natural language expressions, such that a human who reads (and trusts) the first element of a pair would most likely infer that the other element is also true. Paraphrasing can be seen as bidirec- tional textual entailment and methods from the two areas are often very similar. Both kinds of methods are useful, at least in principle, in a wide range of natural language processing applications, including question answering, summarization, text generation, and machine translation. In this thesis, we focus on paraphrase and textual entailment recognition, as well as paraphrase generation. We propose three paraphrase and textual entailment recog- nition methods, experimentally evaluated on existing benchmarks. The key idea is that by capturing similarities at various abstractions of the inputs, we can recognize para- phrases and textual entailment reasonably well. Additionally, we exploit WordNet and use features that operate on the syntactic level of the language expressions. The best of our three recognition methods achieves state of the art results on the widely used MSR paraphrasing corpus, but the simplest of our methods is also a very competitive baseline. On textual entailment datasets, our methods achieve worse results. Neverthe- less, they perform reasonably well, despite being simpler than several other proposed methods; therefore, they can be considered as competitive baselines for future work. ii ABSTRACT iii On the generation side, we propose a novel approach to paraphrasing sentences that produces candidate paraphrases by applying paraphrasing rules and then ranks the candidates. Furthermore, we propose a new methodology to evaluate the ranking com- ponents of generate-and-rank paraphrase generators. A new paraphrasing dataset was also created for evaluations of this kind, which was then used to show that the perfor- mance of our method’s ranker improves when features from our paraphrase recognizer are also included. Furthermore, we show that our method compares well against a state of the art paraphrase generator, and we present evidence that our method might benefit from additional training data. Finally, we present three in vitro studies of how para- phrase generation can be used in question answering, natural language generation, and Web advertisements. Acknowledgements I would like to thank my supervisor, Ion Androutsopoulos, as well as my co-supervisors, Theodore Kalampoukis and Michalis Vazirgiannis, for their guidance throughout my studies at Athens University of Economics and Business (AUEB). The work of this thesis was supported by a scholarship from the Greek PENED 2003 programme, which was co-funded by the European Union (80%) and the Greek General Secretariat for Research and Technology (20%). I would also like to thank all the members of the Natural Language Processing Group of AUEB’s Department of Informatics, especially Dimitris Galanis and Makis Lampouras for the long discussions we had concerning the problems encountered dur- ing my research. A big thanks goes to my family and all my friends for their support and patience all these years. Especially to Matina Charami, who was always there when I needed her, to support me and help me go through many difficulties. Last but not least, I would like to thank my three dogs, Liza, Dahlia, and Phantom, for the moments of joy they have given me, lifting up my spirit when needed. iv Contents Abstract ii Acknowledgements iv Contents v List of Tables viii List of Figures x 1 Introduction 1 1.1 Paraphrasing and textual entailment . 1 1.2 Contribution of this thesis . 3 1.2.1 Extensive survey and classification of previous methods . 3 1.2.2 Paraphrase and textual entailment recognition . 4 1.2.3 Paraphrase generation . 5 1.2.4 Applications of paraphrase generation . 6 1.3 Outline of the remainder of this thesis . 6 2 Background and related work 7 2.1 Introduction . 7 2.2 Paraphrase and textual entailment recognition . 16 2.3 Paraphrase and textual entailment generation . 27 v CONTENTS vi 2.4 Paraphrase and textual entailment extraction . 37 2.5 Conclusions . 51 3 Paraphrase and Textual Entailment Recognition 53 3.1 Introduction . 53 3.2 Datasets . 55 3.3 Evaluation measures . 59 3.4 Our recognition methods . 60 3.4.1 Maximum Entropy classifier . 61 3.4.2 String similarity measures . 63 3.4.3 Combining similarity measures . 65 3.4.4 The need for synonyms and grammatical relations . 70 3.4.5 Feature selection . 74 3.4.6 Baselines . 76 3.5 Comparing to other methods . 78 3.6 Conclusions and contribution of this chapter . 82 4 Paraphrase Generation 84 4.1 Introduction . 84 4.2 Generating candidate paraphrases . 87 4.3 A dataset of sources and candidates . 89 4.4 Quick introduction to SVMs and SVRs .................. 93 4.4.1 Binary classification with SVMs . 94 4.4.2 Ranking with SVRs........................ 96 4.5 Ranking candidate paraphrases . 97 4.6 Comparing to the state of the art . 102 4.7 Learning rate . 106 4.8 Conclusions and contribution of this chapter . 107 CONTENTS vii 5 Applications of Paraphrase Generation 109 5.1 Introduction . 109 5.2 Paraphrasing for question answering . 111 5.2.1 Architecture of QA systems . 111 5.2.2 Paraphrasing questions . 113 5.3 Paraphrasing for natural language generation . 115 5.3.1 Architecture of NLG systems . 115 5.3.2 Paraphrasing sentence templates . 117 5.4 Paraphrasing advertisements . 119 5.5 Conclusions and contribution of this chapter . 123 6 Conclusions 125 6.1 Summary and contribution of this thesis . 125 6.1.1 Extensive survey and classification of previous methods . 125 6.1.2 Paraphrase and textual entailment recognition . 126 6.1.3 Paraphrase generation . 127 6.1.4 Applications of paraphrase generation . 128 6.2 Future work . 128 A Recognition results on all datasets 130 Bibliography 136 List of Tables 3.1 BASE1 results per corpus. 77 3.2 BASE2 results per corpus. 78 3.3 Paraphrase recognition results on the MSR corpus. 79 3.4 Textual entailment recognition results on the RTE-1 corpus. 79 3.5 Textual entailment recognition results on the RTE-2 corpus. 79 3.6 Textual entailment recognition results (for two classes) on the RTE-3 corpus. 80 3.7 Textual entailment recognition results (for three classes) on the RTE-3 corpus. 80 3.8 Textual entailment recognition results (for two classes) on the RTE-4 corpus. 80 3.9 Textual entailment recognition results (for three classes) on the RTE-4 corpus. 81 3.10 Textual entailment recognition results (for two classes) on the RTE-5 corpus. 81 3.11 Textual entailment recognition results (for three classes) on the RTE-5 corpus. 81 4.1 Evaluation guidelines. 91 4.2 Example calculation of k statistic . 92 viii LIST OF TABLES ix 4.3 Inter-annotator agreement. 93 4.4 Comparison of SVR-REC and ZHAO-ENG. 105 A.1 INIT results per corpus. 133 features are used. 131 A.2 INIT+WN results per corpus. 133 features are used. 131 A.3 INIT+WN+DEP results per corpus. 136 features are used. 131 A.4 FHC results per corpus. 132 A.5 FBS results per corpus. 132 A.6 BHC results per corpus. 132 A.7 BBS results per corpus. 133 A.8 Results on MSR corpus. 133 A.9 Results on RTE-1 dataset. 133 A.10 Results on RTE-2 dataset. 134 A.11 Results on RTE-3 dataset. 134 A.12 Results on RTE-4 dataset. 134 A.13 Results on RTE-5 dataset. 135 List of Figures 2.1 Two sentences that are very similar when viewed at the level of depen- dency trees. 18 2.2 An example of how dependency trees may make it easier to match a short sentence (subtree inside the dashed line) to a part of a longer one. 23 2.3 Paraphrase and textual entailment recognition via supervised machine learning. 24 2.4 Generating paraphrases of “X wrote Y” by bootstrapping. 36 2.5 Dependency trees of sentences (2.54) and (2.57). 39 2.6 Word lattices obtained from sentence clusters in Barzilay and Lee’s method. 42 2.7 Merging parse trees of aligned sentences in Pang et al.’s method. 43 2.8 Finite state automaton produced by Pang et al.’s method. 44 3.1 INIT error rate learning curves on the MSR corpus. 68 3.2 INIT error rate learning curves on the RTE-3 corpus for 2 classes. 68 3.3 INIT error rate learning curves on the RTE-3 corpus for 3 classes. 69 3.4 Grammatical relations of example 3.22–3.23. 72 3.5 Grammatical relations of example 3.24–3.25. 73 3.6 INIT+WN+DEP and INIT error rate learning curves on the MSR corpus. 73 3.7 BBS error rate learning curves on the RTE-3 corpus for 2 classes. 75 x LIST OF FIGURES xi 3.8 FHC error rate learning curves on the RTE-3 corpus for 3 classes. 76 4.1 Distribution of overall paraphrase quality scores in the evaluation dataset. 90 4.2 Comparison of SVR-REC and SVR-BASE. 100 4.3 Grammaticality of SVR-REC and ZHAO-ENG. 104 4.4 Meaning preserv. of SVR-REC and ZHAO-ENG. 104 4.5 Diversity of SVR-REC and ZHAO-ENG. 105 4.6 Learning curves of SVR-REC-BIN and ME-REC. 107 5.1 A typical QA system architecture.
Recommended publications
  • Are Annotators' Word-Sense-Disambiguation
    J. Hlavácovᡠ(Ed.): ITAT 2017 Proceedings, pp. 5–14 CEUR Workshop Proceedings Vol. 1885, ISSN 1613-0073, c 2017 S. Cinková, A. Vernerová Are Annotators’ Word-Sense-Disambiguation Decisions Affected by Textual Entailment between Lexicon Glosses? Silvie Cinková, Anna Vernerová Charles University, Faculty of Mathematics and Physics Institute of Formal and Applied Linguistics Malostranské námestíˇ 25 118 00 Praha 1 Czech Republic ufal.mff.cuni.cz {cinkova,vernerova}@ufal.mff.cuni.cz Abstract: We describe an annotation experiment com- verb on the one hand, and those related to the lexicograph- bining topics from lexicography and Word Sense Disam- ical design of PDEV, such as semantic relations between biguation. It involves a lexicon (Pattern Dictionary of En- implicatures within a lemma or denotative similarity of the glish Verbs, PDEV), an existing data set (VPS-GradeUp), verb arguments, on the other hand (see Section 3 for defi- and an unpublished data set (RTE in PDEV Implicatures). nitions and examples). The aim of the experiment was twofold: a pilot annota- This paper focuses on a feature related to PDEV’s de- tion of Recognizing Textual Entailment (RTE) on PDEV sign (see Section 2.3), namely on textual entailment be- implicatures (lexicon glosses) on the one hand, and, on tween implicatures in pairs of patterns of the same lemma the other hand, an analysis of the effect of Textual Entail- entry (henceforth colempats, see Section 3.1 for definition ment between lexicon glosses on annotators’ Word-Sense- and more detail). Disambiguation decisions, compared to other predictors, We pairwise compare all colempats, examining their such as finiteness of the target verb, the explicit presence scores in the graded decision annotation with respect to of its relevant arguments, and the semantic distance be- how much they compete to become the most appropriate tween corresponding syntactic arguments in two different pattern, as well as the scores of presence of textual entail- patterns (dictionary senses).
    [Show full text]
  • Textual Inference for Machine Comprehension Martin Gleize
    Textual Inference for Machine Comprehension Martin Gleize To cite this version: Martin Gleize. Textual Inference for Machine Comprehension. Computation and Language [cs.CL]. Université Paris Saclay (COmUE), 2016. English. NNT : 2016SACLS004. tel-01317577 HAL Id: tel-01317577 https://tel.archives-ouvertes.fr/tel-01317577 Submitted on 18 May 2016 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Le cas échéant, logo de l’établissement co-délivrant le doctorat en cotutelle internationale de thèse , sinon mettre le logo de l’établissement de préparation de la thèse (UPSud, HEC, UVSQ, UEVE, ENS Cachan, Polytechnique, IOGS, …) NNT : 2016SACLS004 THÈSE DE DOCTORAT DE L’UNIVERSITÉ PARIS-SACLAY PRÉPARÉE À L'UNIVERSITÉ PARIS-SUD ECOLE DOCTORALE N° 580 Sciences et technologies de l'information et de la communication (STIC) Spécialité de doctorat : Informatique Par M. Martin Gleize Textual Inference for Machine Comprehension Thèse présentée et soutenue à Orsay, le 7 janvier 2016 Composition du Jury : M. Yvon François Professeur, Université Paris-Sud Président, Examinateur Mme Gardent Claire Directeur de recherche CNRS, LORIA Rapporteur M. Magnini Bernardo Senior Researcher, FBK Rapporteur M. Piwowarski Benjamin Chargé de recherche CNRS, LIP6 Examinateur Mme Grau Brigitte Professeur, ENSIIE Directeur de thèse Abstract With the ever-growing mass of published text, natural language understanding stands as one of the most sought-after goal of artificial intelligence.
    [Show full text]
  • Repurposing Entailment for Multi-Hop Question Answering Tasks
    Repurposing Entailment for Multi-Hop Question Answering Tasks Harsh Trivedi|, Heeyoung Kwon|, Tushar Khot♠, Ashish Sabharwal♠, Niranjan Balasubramanian| | Stony Brook University, Stony Brook, U.S.A. fhjtrivedi,heekwon,[email protected] ♠ Allen Institute for Artificial Intelligence, Seattle, U.S.A. ftushark,[email protected] Abstract Question Answering (QA) naturally reduces to an entailment problem, namely, verifying whether some text entails the answer to a ques- tion. However, for multi-hop QA tasks, which require reasoning with multiple sentences, it remains unclear how best to utilize entailment models pre-trained on large scale datasets such as SNLI, which are based on sentence pairs. We introduce Multee, a general architecture Figure 1: An example illustrating the challenges in us- that can effectively use entailment models for ing sentence-level entailment model for multi-sentence multi-hop QA tasks. Multee uses (i) a local reasoning needed for QA, and the high-level approach module that helps locate important sentences, used in Multee. thereby avoiding distracting information, and level, whereas question answering requires verify- (ii) a global module that aggregates informa- ing whether multiple sentences, taken together as tion by effectively incorporating importance weights. Importantly, we show that both mod- a premise, entail a hypothesis. ules can use entailment functions pre-trained There are two straightforward ways to address on a large scale NLI datasets. We evaluate per- this mismatch: (1) aggregate independent entail- formance on MultiRC and OpenBookQA, two ment decisions over each premise sentence, or (2) multihop QA datasets. When using an entail- make a single entailment decision after concate- ment function pre-trained on NLI datasets, nating all premise sentences.
    [Show full text]
  • Techniques for Recognizing Textual Entailment and Semantic Equivalence
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by EPrints Complutense Techniques for Recognizing Textual Entailment and Semantic Equivalence Jes´us Herrera, Anselmo Pe˜nas, Felisa Verdejo Departamento de Lenguajes y Sistemas Inform´aticos Universidad Nacional de Educaci´ona Distancia Madrid, Spain {jesus.herrera, anselmo, felisa}@lsi.uned.es Abstract. After defining what is understood by textual entailment and semantic equivalence, the present state and the desirable future of the systems aimed at recognizing them is shown. A compilation of the currently implemented techniques in the main Recognizing Textual Entailment and Semantic Equivalence systems is given. 1 Introduction The concept “textual entailment” is used to indicate the state in which the semantics of a natural language written text can be inferred from the semantics of another one. More specifically, if the truth of an enunciation entails the truth of another enunciation. For example, given the texts: 1. The three-day G8 meeting will take place in Scotland. 2. The Group of Eight summit will last three days. it is clear that the semantics of the second one can be inferred from the semantics of the first one; then, it is said that textual entailment exists between both texts. Textual entailment is a directional relationship: in the example above, the first statement entails the second one, but this entailment is not given in the opposite direction. The recognition of textual entailment requires a processing at the lexical level (for example, synonymy between meeting and summit or between G8 and Group of Eight), as well as at the syntactic level and the sentence semantic level.
    [Show full text]
  • Automatic Evaluation of Summary Using Textual Entailment
    Automatic Evaluation of Summary Using Textual Entailment Pinaki Bhaskar Partha Pakray Department of Computer Science & Department of Computer Science & Engineering, Jadavpur University, Engineering, Jadavpur University, Kolkata – 700032, India Kolkata – 700032, India [email protected] [email protected] summary is very much needed when a large Abstract number of summaries have to be evaluated, spe- cially for multi-document summaries. For sum- This paper describes about an automatic tech- mary evaluation we have developed an automat- nique of evaluating summary. The standard ed evaluation technique based on textual entail- and popular summary evaluation techniques or ment. tools are not fully automatic; they all need Recognizing Textual Entailment (RTE) is one some manual process. Using textual entail- of the recent research areas of Natural Language ment (TE) the generated summary can be evaluated automatically without any manual Processing (NLP). Textual Entailment is defined evaluation/process. The TE system is the as a directional relationship between pairs of text composition of lexical entailment module, lex- expressions, denoted by the entailing “Text” (T) ical distance module, Chunk module, Named and the entailed “Hypothesis” (H). T entails H if Entity module and syntactic text entailment the meaning of H can be inferred from the mean- (TE) module. The syntactic TE system is ing of T. Textual Entailment has many applica- based on the Support Vector Machine (SVM) tions in NLP tasks, such as Summarization, In- that uses twenty five features for lexical simi- formation Extraction, Question Answering, In- larity, the output tag from a rule based syntac- formation Retrieval. tic two-way TE system as a feature and the outputs from a rule based Chunk Module and 2 Related Work Named Entity Module as the other features.
    [Show full text]
  • Dependency-Based Textual Entailment
    Dependency-based Textual Entailment Vasile Rus Department of Computer Science Institute for Intelligent Systems The University of Memphis Memphis, TN 38120, USA [email protected] Abstract and Hypothesis are mapped into two sets of dependen- cies: the T-set and H-set, respectively. A score is then This paper studies the role of dependency information computed that measures the degree to which the depen- for the task of textual entailment. Both the Text and Hypothesis of an entailment pair are mapped into sets dencies in the H-set are present in the T-set. Based on of dependencies and a score is computed that measures the score, an entailment decision is made. Dependen- the similarity of the two sets. Based on the score an cies are labelled relations among words in a sentence. entailment decision is made. Two experiments are con- For example, there is a subj dependency between cap- ducted to measure the impact of dependencies on the ital and Besancon in the previous example which we entailment task. In one experiment we compare the represent as the triplet (subj, capital, Besancon). The dependency-based approach with a baseline approach first term in the triplet is the name or label of the re- on a standard data set. In a second experiment, we lation, the second term represents the head word, and measure the performance on a subset of the standard the last term is the dependent word. We use MINIPAR data set. The subset is so selected to minimize the (Lin 1998), an automated dependency parser, as our effect of other factors, such as word-level information, in the performance measurement process.
    [Show full text]
  • Parser Evaluation Using Textual Entailments
    Language Resources and Evaluation manuscript No. (will be inserted by the editor) Parser Evaluation Using Textual Entailments Deniz Yuret · Laura Rimell · Aydın Han Received: date / Accepted: date Abstract Parser Evaluation using Textual Entailments (PETE) is a shared task in the SemEval-2010 Evaluation Exercises on Semantic Evaluation. The task involves recognizing textual entailments based on syntactic information alone. PETE introduces a new parser evaluation scheme that is formalism independent, less prone to annotation error, and focused on semantically rele- vant distinctions. This paper describes the PETE task, gives an error analysis of the top-performing Cambridge system, and introduces a standard entail- ment module that can be used with any parser that outputs Stanford typed dependencies. Keywords Parsing · Textual Entailments 1 Introduction Parser Evaluation using Textual Entailments (PETE) is a shared task in the SemEval-2010 Evaluation Exercises on Semantic Evaluation that involves rec- ognizing textual entailments to evaluate parser performance. Given two text Deniz Yuret and Aydın Han Ko¸cUniversity Rumelifeneri Yolu 34450 Sarıyer, Istanbul,˙ Turkey Tel.: +90-212-338-1724 Fax: +90-212-338-1548 E-mail: dyuret,[email protected] Laura Rimell Computer Laboratory William Gates Building 15 JJ Thomson Avenue Cambridge CB3 0FD, UK Tel.: +44 (0)1223 334696 E-mail: [email protected] 2 Deniz Yuret et al. fragments called “text” (T) and “hypothesis” (H), recognizing textual entail- ment (RTE) is the task of determining whether the meaning of the hypothesis is entailed (can be inferred) from the text. In contrast to general RTE tasks (Dagan et al 2009) PETE is a targeted textual entailment task that focuses on syntactic entailments: Text: The man with the hat was tired.
    [Show full text]
  • Textual Entailment Recognition Based on Dependency Analysis and Wordnet
    Textual Entailment Recognition Based on Dependency Analysis and WordNet Jesus´ Herrera, Anselmo Penas,˜ Felisa Verdejo Departamento de Lenguajes y Sistemas Informaticos´ Universidad Nacional de Educacion´ a Distancia Madrid, Spain jesus.herrera, anselmo, felisa @lsi.uned.es f g Abstract. The Recognizing Textual Entailment System shown here is based on the use of a broad-coverage parser to extract dependency relationships; in addi- tion, WordNet relations are used to recognize entailment at the lexical level. The work investigates whether the mapping of dependency trees from text and hy- pothesis give better evidence of entailment than the matching of plain text alone. While the use of WordNet seems to improve system's performance, the notion of mapping between trees here explored (inclusion) shows no improvement, sug- gesting that other notions of tree mappings should be explored such as tree edit distances or tree alignment distances. 1 Introduction Textual Entailment Recognition (RTE) aims at deciding whether the truth of a text en- tails the truth of another text called hypothesis. This concept has been the basis for the PASCAL1 RTE Challenge [3]. The system presented here is aimed at validating the hypothesis that (i) a certain amount of semantic information could be extracted from texts by means of the syntactic structure given by a dependency analysis, and that (ii) lexico-semantic information such as WordNet relations can improve RTE. In short, the techniques involved in this system are the following: – Dependency analysis of texts and hypothesises. – Lexical entailment between dependency tree nodes using WordNet. – Mapping between dependency trees based on the notion of inclusion.
    [Show full text]
  • A Probabilistic Classification Approach for Lexical Textual Entailment
    A Probabilistic Classification Approach for Lexical Textual Entailment Oren Glickman and Ido Dagan and Moshe Koppel Computer Science Department , Bar Ilan University Ramat-Gan, 52900, Iarael {glikmao, dagan, koppel}@cs.biu.ac.il Abstract The uncertain nature of textual entailment calls for its The textual entailment task – determining if a given text explicit modeling in probabilistic terms. In this paper we entails a given hypothesis – provides an abstraction of first propose a general generative probabilistic setting for applied semantic inference. This paper describes first a textual entailment. This probabilistic framework may be general generative probabilistic setting for textual considered analogous to (though different than) the entailment. We then focus on the sub-task of recognizing probabilistic setting defined for other phenomena, like whether the lexical concepts present in the hypothesis are language modeling and statistical machine translation. entailed from the text. This problem is recast as one of text Obviously, such frameworks are needed to allow for categorization in which the classes are the vocabulary principled (fuzzy) modeling of the relevant language words. We make novel use of Naïve Bayes to model the problem in an entirely unsupervised fashion. Empirical tests phenomenon, compared to utilizing ad-hoc ranking scores suggest that the method is effective and compares favorably which have no clear probabilistic interpretation. We with state-of-the-art heuristic scoring approaches. suggest that the proposed setting may provide a unifying framework for modeling uncertain semantic inferences from texts. Introduction An important sub task of textual entailment, which we term lexical entailment , is recognizing if the concepts in a Many Natural Language Processing (NLP) applications hypothesis h are entailed from a given text t, even if the need to recognize when the meaning of one text can be relations between these concepts may not be entailed from expressed by, or inferred from, another text.
    [Show full text]
  • Frame Semantics Annotation Made Easy with Dbpedia
    Frame Semantics Annotation Made Easy with DBpedia Marco Fossati, Sara Tonelli, and Claudio Giuliano ffossati, satonelli, [email protected] Fondazione Bruno Kessler - via Sommarive, 18 - 38123 Trento, Italy Abstract. Crowdsourcing techniques applied to natural language pro- cessing have recently experienced a steady growth and represent a cheap and fast, albeit valid, solution to create benchmarks and training data. Nevertheless, some particularly complex tasks such as semantic role an- notation have been rarely conducted in a crowdsourcing environment, due to their intrinsic difficulty. In this paper, we present a novel ap- proach to accomplish this task by leveraging information automatically extracted from DBpedia. We show that replacing role definitions, typi- cally meant for expert annotators, with a list of DBpedia types, makes the identification and assignment of role labels more intuitive also for non-expert workers. Results prove that such strategy improves on the standard annotation workflow, both in terms of accuracy and of time consumption. Keywords: Natural Language Processing, Frame Semantics, En- tity Linking, DBpedia, Crowdsourcing, Task Modeling 1 Introduction Frame semantics [6] is one of the theories that originate from the long strand of linguistic research in artificial intelligence. A frame can be informally defined as an event triggered by some term in a text and embedding a set of partic- ipants. For instance, the sentence Goofy has murdered Mickey Mouse evokes the Killing frame (triggered by murdered) together with the Killer and Vic- tim participants (respectively Goofy and Mickey Mouse). Such theory has led to the creation of FrameNet [2], namely an English lexical database containing manually annotated textual examples of frame usage.
    [Show full text]
  • Constructing a Textual Semantic Relation Corpus Using a Discourse Treebank
    Constructing a Textual Semantic Relation Corpus Using a Discourse Treebank Rui Wang and Caroline Sporleder Computational Linguistics Saarland University frwang,[email protected] Abstract In this paper, we present our work on constructing a textual semantic relation corpus by making use of an existing treebank annotated with discourse relations. We extract adjacent text span pairs and group them into six categories according to the different discourse relations between them. After that, we present the details of our annotation scheme, which includes six textual semantic relations, backward entailment, forward entailment, equality, contradiction, overlapping, and independent. We also discuss some ambiguous examples to show the difficulty of such annotation task, which cannot be easily done by an automatic mapping between discourse relations and semantic relations. We have two annotators and each of them performs the task twice. The basic statistics on the constructed corpus looks promising: we achieve 81.17% of agreement on the six semantic relation annotation with a .718 kappa score, and it increases to 91.21% if we collapse the last two labels with a .775 kappa score. 1. Introduction TRADICTION, and UNKNOWN. ENTAILMENT is the same Traditionally, the term ‘semantic relation’ refers to rela- as the previous YES; but NO is divided into CONTRADIC- tions that hold between the meanings of two words, e.g., TION and UNKNOWN, to differentiate cases where T and H synonymy, hypernymy, etc. These relations are usually are contradictory to each other from all the other cases. The situation-independent. Discourse relations, on the other RTE data are acquired from other natural language process- hand, tend to depend on the wider context of an utterance.
    [Show full text]
  • PATTY: a Taxonomy of Relational Patterns with Semantic Types
    PATTY: A Taxonomy of Relational Patterns with Semantic Types Ndapandula Nakashole, Gerhard Weikum, Fabian Suchanek Max Planck Institute for Informatics Saarbrucken,¨ Germany {nnakasho,weikum,suchanek}@mpi-inf.mpg.de Abstract Yet, existing large-scale knowledge bases are mostly limited to abstract binary relationships be- This paper presents PATTY: a large resource for textual patterns that denote binary relations tween entities, such as “bornIn” (Auer 2007; Bol- between entities. The patterns are semanti- lacker 2008; Nastase 2010; Suchanek 2007). These cally typed and organized into a subsumption do not correspond to real text phrases. Only the Re- taxonomy. The PATTY system is based on ef- Verb system (Fader 2011) yields a larger number of ficient algorithms for frequent itemset mining relational textual patterns. However, no attempt is and can process Web-scale corpora. It har- made to organize these patterns into synonymous nesses the rich type system and entity popu- patterns, let alone into a taxonomy. Thus, the pat- lation of large knowledge bases. The PATTY taxonomy comprises 350,569 pattern synsets. terns themselves do not exhibit semantics. Random-sampling-based evaluation shows a Goal. Our goal in this paper is to systematically pattern accuracy of 84.7%. PATTY has 8,162 compile relational patterns from a corpus, and to im- subsumptions, with a random-sampling-based pose a semantically typed structure on them. The precision of 75%. The PATTY resource is freely available for interactive access and result we aim at is a WordNet-style taxonomy of download. binary relations. In particular, we aim at patterns that contain semantic types, such as hsingeri sings 1 Introduction hsongi.
    [Show full text]