Sequential Tagging of Semantic Roles on Chinese Framenet

Total Page:16

File Type:pdf, Size:1020Kb

Sequential Tagging of Semantic Roles on Chinese Framenet Sequential Tagging of Semantic Roles on Chinese FrameNet Jihong LI Ruibo WANG, Yahui GAO Computer Center Computer Center Shanxi University Shanxi University [email protected] wangruibo,gaoyahui @sxu.edu.cn { } Abstract FrameNet (Baker et al., 1998), PropBank (Kings- bury and Palmer, 2002), NomBank (Meyers et al., In this paper, semantic role labeling(SRL) 2004), and so on. On this basis, several interna- on Chinese FrameNet is divided into the tional semantic evaluations have been organized, subtasks of boundary identification(BI) which include Senseval 3 (Litkowski, 2004), and semantic role classification(SRC). SemEval 2007 (Baker,et al., 2007), CoNLL These subtasks are regarded as the se- 2008 (Surdeanu et al., 2008), CoNLL 2009 (Hajic quential tagging problem at the word et al., 2009), and so on. level, respectively. We use the conditional The first SRL model on FrameNet was pro- random fields(CRFs) model to train and posed by Gildea and Jurafsky(2002). The model test on a two-fold cross-validation data consists of two subtasks of boundary identifica- set. The extracted features include 11 tion(BI) and semantic role classification(SRC). word-level and 15 shallow syntactic fea- Both subtasks were implemented on the pretreat- tures derived from automatic base chunk ment results of the full parsing tree. Many lex- parsing. We use the orthogonal array of ical and syntactic features were extracted to im- statistics to arrange the experiment so that prove the accuracy of the model. On the test data the best feature template is selected. The of FrameNet, the system achieved 65% precision experimental results show that given the and 61% recall. target word within a sentence, the best Most works on SRL followed Gildea’s frame- F-measures of SRL can achieve 60.42%. work of processing the SRL task on English For the BI and SRC subtasks, the best F- FrameNet and PropBank. They built their model measures are 70.55 and 81%, respectively. on the full parse tree and selected features using The statistical t-test shows that the im- various machine learning methods to improve the provement of our SRL model is not signif- accuracy of SRL models. Many attempts have icant after appending the base chunk fea- made significant progress, ssuch as the works of tures. Pradhan et al. (2005), Surdeanu et al. (2007), and so on. Other researchers regarded the task 1 Introduction of SRL as a sequential tagging problem and em- Semantic parsing is important in natural lan- ployed the shallow chunking technique to solve it, guage processing, and it has attracted an increas- as described by Marquez at al. (2008). ing number of studies in recent years. Cur- Although the SRL model based on a full parse rently, its most important aspect is the formaliza- tree has good performance in English, this method tion of the proposition meaning of one sentence of processing is not available in other languages, through the semantic role labeling. Therefore, especially in Chinese. A systemic study of Chi- many large human-annotated corpora have been nese SRL was done by Xue et al. (2008). Like constructed to support related research, such as the English SRL procedure, he removed many 22 Proceedings of the 8th Workshop on Asian Language Resources, pages 22–29, Beijing, China, 21-22 August 2010. c 2010 Asian Federation for Natural Language Processing uncorrelated constituents of a parse tree and re- a sequential tagging problem at the word level. lied on the remainder to identify the semantic Conditional random fields(CRFs) model is em- roles using the maximum entropy model. When ployed to train the model and predict the result human-corrected parse is used, the F-measures of the unlabeled sentence. To improve the accu- on the PropBank and NomBank achieve 92.0 and racy of the model, base chunk features are intro- 69.6%, respectively. However, when automatic duced, and the feature selection method involving full parse is used, the F-measures only achieve an orthogonal array is adopted. The experimen- 71.9 and 60.4%, respectively. This significant de- tal results illustrate that the F-measure of our SRL crease prompts us to analyze its causes and to find model achieves 60.42%. This is the best SRL re- a potential solution. sult of CFN so far. First, the Chinese human-annotated resources The paper is organized as follows. In Section of semantic roles are relatively small. Sun and 2, we describe the situation of CFN and introduce Gildea only studied the SRL of 10 Chinese verbs SRL on CFN. In Section 3, we propose our SRL and extracted 1,138 sentences in the Chinese Tree model in detail. In Section 4, the candidate feature Bank. The size of the Chinese PropBank and set is proposed, and the orthogonal-array-based Chinese NomBank used in the paper of Xue is feature selection method is introduced. In Sec- significantly smaller than the ones used in En- tion 5, we describe the experimental setup used glish language studies. Moreover, more verbs ex- throughout this paper. In Section 6, we list our ist in Chinese than in English, which increases the experimental results and provide detailed analy- sparsity of Chinese Semantic Role data resources. sis. The conclusions and several further directions The same problem also exists in our experiment. are given at the end of this paper. The current corpus of Chinese FrameNet includes about 18,322 human-annotated sentences of 1,671 2 CFN and Its SRL task target words. There is only an average of less than Chinese FrameNet(CFN) (You et al., 2005) is a re- 10 sentences for every target word. To reduce the search project that has been developed by Shanxi influence of the data sparsity, we adopt a two-fold University, creating an FN-styled lexicon for Chi- cross validation technique for train and test label- nese, based on the theory of Frame Semantics ing. (Fillmore, 1982) and supported by corpus evi- Second, because of the lack of morphological dence. The results of the CFN project include a clues in Chinese, the accuracy of a state-of-the-art lexical resource, called the CFN database, and as- parsing system significantly decreases when used sociated software tools. Many natural language for a realistic scenario. In the preliminary stage processing(NLP) applications, such as Informa- of building an SRL model of CFN, we employed tion Retrieval and Machine Translation, will ben- a Stanford full parser to parse all sentences in the efit from this resource. In FN, the semantic roles corpus and adopted the traditional SRL technique of a predicate are called the frame elements of a on our data set. However, the experiment result frame. A frame has different frame elements. A was insignificant. Only 76.48% of the semantic group of lexical units (LUs) that evokes the same roles in the data set have a constituent with the frame share the same names of frame elements. same text span in the parse tree, and the F-measure The CFN project currently contains more than of BI can only achieves 54%. Therefore, we at- 1,671 LUs, more than 219 semantic frames, and tempted to use another processing technique for has exemplified more than 18,322 annotated sen- SRL on CFN. We formalized SRL on CFN into a tences. In addition to correct segmentation and sequential tagging problem at the word level. We part of speech, every sentence in the database is first extracted 11 word features into the baseline marked up to exemplify the semantic and syntac- model. Then we added 15 additional base chunk tic information of the target word. Each annotated features into the SRL model. sentence contains only one target word. In this paper, the SRL task of CFN comprises (a). <medium-np-subj 第 1/m 章/q > <tgt=” two subtasks: BI and SRC. These are regarded as 陈述” 介绍/v > <msg-np-obj 算法/n 与/c 数据/n 23 结构/n > ;/w To avoid the problem of data sparsity, we use all The CFN Corpus is currently at an early stage, sentences in our train data set to train the model of and the available CFN resource is relatively lim- BI. ited, so the SRL task on CFN is described as fol- lows. Given a Chinese sentence, a target word, and its frame, we identify the boundaries of the 3.3 SRC frame elements within the sentence and label them with the appropriate frame element name. This is After predicting the boundaries of semantic role the same as the task in Senseval-3. chunks in a sentence, the proper semantic role types should be assigned in the SRC step. Al- 3 Shallow SRL Models though it can be easily modeled as a classification problem, we regarded it as a sequential tagging This section proposes our SRL model architec- problem at the word level. An additional con- ture, and describes the stages of our model in de- straint is employed in this step: the boundary tags tail. of the predicting sequence of this stage should be 3.1 SRL Model Architecture consistent with the the output of the BI stage. A family of SRL models can be constructed using One intuitive reason for this model strategy is only shallow syntactic information as the input. that the SRC step can use the same feature set as The main differences of the models in this family BI, and it can further prove the rationality of our mainly focus on the following two aspects. feature optimization method. i) model strategy: whether to combine the sub- tasks of BI and SRC? ii) tagging unit: which is used as the tagging 3.4 Postprocessing unit, word or chunk.
Recommended publications
  • Enriching an Explanatory Dictionary with Framenet and Propbank Corpus Examples
    Proceedings of eLex 2019 Enriching an Explanatory Dictionary with FrameNet and PropBank Corpus Examples Pēteris Paikens 1, Normunds Grūzītis 2, Laura Rituma 2, Gunta Nešpore 2, Viktors Lipskis 2, Lauma Pretkalniņa2, Andrejs Spektors 2 1 Latvian Information Agency LETA, Marijas street 2, Riga, Latvia 2 Institute of Mathematics and Computer Science, University of Latvia, Raina blvd. 29, Riga, Latvia E-mail: [email protected], [email protected] Abstract This paper describes ongoing work to extend an online dictionary of Latvian – Tezaurs.lv – with representative semantically annotated corpus examples according to the FrameNet and PropBank methodologies and word sense inventories. Tezaurs.lv is one of the largest open lexical resources for Latvian, combining information from more than 300 legacy dictionaries and other sources. The corpus examples are extracted from Latvian FrameNet and PropBank corpora, which are manually annotated parallel subsets of a balanced text corpus of contemporary Latvian. The proposed approach augments traditional lexicographic information with modern cross-lingually interpretable information and enables analysis of word senses from the perspective of frame semantics, which is substantially different from (complementary to) the traditional approach applied in Latvian lexicography. In cases where FrameNet and PropBank corpus evidence aligns well with the word sense split in legacy dictionaries, the frame-semantically annotated corpus examples supplement the word sense information with clarifying usage examples and commonly used semantic valence patterns. However, the annotated corpus examples often provide evidence of a different sense split, which is often more coarse-grained and, thus, may help dictionary users to cluster and comprehend a fine-grained sense split suggested by the legacy sources.
    [Show full text]
  • Evaluation of Machine Learning Algorithms for Sms Spam Filtering
    EVALUATION OF MACHINE LEARNING ALGORITHMS FOR SMS SPAM FILTERING David Bäckman Bachelor Thesis, 15 credits Bachelor Of Science Programme in Computing Science 2019 Abstract The purpose of this thesis is to evaluate dierent machine learning algorithms and methods for text representation in order to determine what is best suited to use to distinguish between spam SMS and legitimate SMS. A data set that contains 5573 real SMS has been used to train the algorithms K-Nearest Neighbor, Support Vector Machine, Naive Bayes and Logistic Regression. The dierent methods that have been used to represent text are Bag of Words, Bigram and Word2Vec. In particular, it has been investigated if semantic text representations can improve the performance of classication. A total of 12 combinations have been evaluated with help of the metrics accuracy and F1-score. The results shows that Logistic Regression together with Bag of Words reach the highest accuracy and F1-score. Bigram as text representation seems to work worse then the others methods. Word2Vec can increase the performnce for K- Nearst Neigbor but not for the other algorithms. Acknowledgements I would like to thank my supervisor Kai-Florian Richter for all good advice and guidance through the project. I would also like to thank all my classmates for help and support during the education, you have made it possible for me to reach this day. Contents 1 Introduction 1 1.1 Background1 1.2 Purpose and Research Questions1 2 Related Work 3 3 Theoretical Background 5 3.1 The Choice of Algorithms5 3.2 Classication
    [Show full text]
  • NLP - Assignment 2
    NLP - Assignment 2 Week 2 December 27th, 2016 1. A 5-gram model is a order Markov Model: (a) Six (b) Five (c) Four (d) Constant Ans : c) Four 2. For the following corpus C1 of 3 sentences, what is the total count of unique bi- grams for which the likelihood will be estimated? Assume we do not perform any pre-processing, and we are using the corpus as given. (i) ice cream tastes better than any other food (ii) ice cream is generally served after the meal (iii) many of us have happy childhood memories linked to ice cream (a) 22 (b) 27 (c) 30 (d) 34 Ans : b) 27 3. Arrange the words \curry, oil and tea" in descending order, based on the frequency of their occurrence in the Google Books n-grams. The Google Books n-gram viewer is available at https://books.google.com/ngrams: (a) tea, oil, curry (c) curry, tea, oil (b) curry, oil, tea (d) oil, tea, curry Ans: d) oil, tea, curry 4. Given a corpus C2, The Maximum Likelihood Estimation (MLE) for the bigram \ice cream" is 0.4 and the count of occurrence of the word \ice" is 310. The likelihood of \ice cream" after applying add-one smoothing is 0:025, for the same corpus C2. What is the vocabulary size of C2: 1 (a) 4390 (b) 4690 (c) 5270 (d) 5550 Ans: b)4690 The Questions from 5 to 10 require you to analyse the data given in the corpus C3, using a programming language of your choice.
    [Show full text]
  • 3 Dictionaries and Tolerant Retrieval
    Online edition (c)2009 Cambridge UP DRAFT! © April 1, 2009 Cambridge University Press. Feedback welcome. 49 Dictionaries and tolerant 3 retrieval In Chapters 1 and 2 we developed the ideas underlying inverted indexes for handling Boolean and proximity queries. Here, we develop techniques that are robust to typographical errors in the query, as well as alternative spellings. In Section 3.1 we develop data structures that help the search for terms in the vocabulary in an inverted index. In Section 3.2 we study WILDCARD QUERY the idea of a wildcard query: a query such as *a*e*i*o*u*, which seeks doc- uments containing any term that includes all the five vowels in sequence. The * symbol indicates any (possibly empty) string of characters. Users pose such queries to a search engine when they are uncertain about how to spell a query term, or seek documents containing variants of a query term; for in- stance, the query automat* would seek documents containing any of the terms automatic, automation and automated. We then turn to other forms of imprecisely posed queries, focusing on spelling errors in Section 3.3. Users make spelling errors either by accident, or because the term they are searching for (e.g., Herman) has no unambiguous spelling in the collection. We detail a number of techniques for correcting spelling errors in queries, one term at a time as well as for an entire string of query terms. Finally, in Section 3.4 we study a method for seeking vo- cabulary terms that are phonetically close to the query term(s).
    [Show full text]
  • Developing a Large Scale Framenet for Italian: the Iframenet Experience
    Developing a large scale FrameNet for Italian: the IFrameNet experience Roberto Basili° Silvia Brambilla§ Danilo Croce° Fabio Tamburini§ ° § Dept. of Enterprise Engineering Dept. of Classic Philology and Italian Studies University of Rome Tor Vergata University of Bologna {basili,croce}@info.uniroma2.it [email protected], [email protected] ian Portuguese, German, Spanish, Japanese, Swe- Abstract dish and Korean. All these projects are based on the idea that English. This paper presents work in pro- most of the Frames are the same among languages gress for the development of IFrameNet, a and that, thanks to this, it is possible to adopt large-scale, computationally oriented, lexi- Berkeley’s Frames and FEs and their relations, cal resource based on Fillmore’s frame se- with few changes, once all the language-specific mantics for Italian. For the development of information has been cut away (Tonelli et al. 2009, IFrameNet linguistic analysis, corpus- Tonelli 2010). processing and machine learning techniques With regard to Italian, over the past ten years are combined in order to support the semi- several research projects have been carried out at automatic development and annotation of different universities and Research Centres. In par- the resource. ticular, the ILC-CNR in Pisa (e.g. Lenci et al. 2008; Johnson and Lenci 2011), FBK in Trento (e.g. Italiano. Questo articolo presenta un work Tonelli et al. 2009, Tonelli 2010) and the Universi- in progress per lo sviluppo di IFrameNet, ty of Rome, Tor Vergata (e.g. Pennacchiotti et al. una risorsa lessicale ad ampia copertura, 2008, Basili et al.
    [Show full text]
  • Integrating Wordnet and Framenet Using a Knowledge-Based Word Sense Disambiguation Algorithm
    Integrating WordNet and FrameNet using a knowledge-based Word Sense Disambiguation algorithm Egoitz Laparra and German Rigau IXA group. University of the Basque Country, Donostia {egoitz.laparra,german.rigau}@ehu.com Abstract coherent groupings of words belonging to the same frame. This paper presents a novel automatic approach to In that way we expect to extend the coverage of FrameNet partially integrate FrameNet and WordNet. In that (by including from WordNet closely related concepts), to way we expect to extend FrameNet coverage, to en- enrich WordNet with frame semantic information (by port- rich WordNet with frame semantic information and ing frame information to WordNet) and possibly to extend possibly to extend FrameNet to languages other than FrameNet to languages other than English (by exploiting English. The method uses a knowledge-based Word local wordnets aligned to the English WordNet). Sense Disambiguation algorithm for linking FrameNet WordNet1 [12] (hereinafter WN) is by far the most lexical units to WordNet synsets. Specifically, we ex- ploit a graph-based Word Sense Disambiguation algo- widely-used knowledge base. In fact, WN is being rithm that uses a large-scale knowledge-base derived used world-wide for anchoring different types of seman- from WordNet. We have developed and tested four tic knowledge including wordnets for languages other than additional versions of this algorithm showing a sub- English [4], domain knowledge [17] or ontologies like stantial improvement over previous results. SUMO [22] or the EuroWordNet Top Concept Ontology [3]. It contains manually coded information about En- glish nouns, verbs, adjectives and adverbs and is organized around the notion of a synset.
    [Show full text]
  • The Hebrew Framenet Project
    The Hebrew FrameNet Project Avi Hayoun, Michael Elhadad Dept. of Computer Science Ben-Gurion University Beer Sheva, Israel {hayounav,elhadad}@cs.bgu.ac.il Abstract We present the Hebrew FrameNet project, describe the development and annotation processes and enumerate the challenges we faced along the way. We have developed semi-automatic tools to help speed the annotation and data collection process. The resource currently covers 167 frames, 3,000 lexical units and about 500 fully annotated sentences. We have started training and testing automatic SRL tools on the seed data. Keywords: FrameNet, Hebrew, frame semantics, semantic resources 1. Introduction frames and their structures in natural language. Recent years have seen growing interest in the task 1.2. FrameNet in Other Languages of Semantic Role Labeling (SRL) of natural language The original FrameNet project has been adapted and text (sometimes called “shallow semantic parsing”). ported to multiple languages. The most active interna- The task is usually described as the act of identify- tional FrameNet teams include the Swedish FrameNet ing the semantic roles, which are the set of semantic (SweFN) covering close to 1,200 frames with 34K LUs properties and relationships defined over constituents (Ahlberg et al., 2014); the Japanese FrameNet (JFN) of a sentence, given a semantic context. with 565 frames, 8,500 LUs and 60K annotated ex- The creation of resources that document the realiza- ample sentences (Ohara, 2013); and FrameNet Brazil tion of semantic roles in natural language texts, such (FN-Br) covering 179 frames, 196 LUs and 12K anno- as FrameNet (Fillmore and Baker, 2010; Ruppenhofer tated sentences (Torrent and Ellsworth, 2013).
    [Show full text]
  • INF5830 Introduction to Semantic Role Labeling
    INF5830 Introduction to Semantic Role Labeling Andrey Kutuzov University of Oslo Language Technology Group With thanks to Lilja Øvrelid, Martha Palmer and Dan Jurafsky INF5830 Introduction to Semantic Role Labeling 1(36) Semantic Role Labeling INF5830 Introduction to Semantic Role Labeling 2(36) Introduction Contents Introduction Semantic roles in general PropBank: Proto-roles FrameNet: Frame Semantics Summary INF5830 Introduction to Semantic Role Labeling 2(36) Introduction Semantics I Study of meaning, expressed in language; I Morphemes, words, phrases, sentences; I Lexical semantics; I Sentence semantics; I (Pragmatics: how the context affects meaning). INF5830 Introduction to Semantic Role Labeling 3(36) Introduction Semantics I Linguistic knowledge: meaning I Meaningful or not: I Word { flick vs blick I Sentence { John swims vs John metaphorically every I Several meanings (WSD): I Words { fish I Sentence { John saw the man with the binoculars I Same meaning (semantic similarity): I Word { sofa vs couch I Sentence { John gave Hannah a gift vs John gave a gift to Hannah I Truth conditions: I All kings are male I Molybdenum conducts electricity I Entailment: I Alfred murdered the librarian I The librarian is dead I Participant roles: John is the `giver', Hannah is the `receiver' INF5830 Introduction to Semantic Role Labeling 4(36) Introduction Representing events I We want to understand the event described by these sentences: 1. IBM bought Spark 2. IBM acquired Spark 3. Spark was acquired by IBM 4. The owners of Spark sold it to IBM I Dependency parsing is insufficient. UDPipe will give us simple relations between verbs and arguments: 1. (buy, nsubj, IBM), (buy, obj, Spark) 2.
    [Show full text]
  • Using Frame Semantics in Natural Language Processing
    Using Frame Semantics in Natural Language Processing Apoorv Agarwal Daniel Bauer Owen Rambow Dept. of Computer Science Dept. of Computer Science CCLS Columbia University Columbia University Columbia University New York, NY New York, NY New York, NY [email protected] [email protected] [email protected] Abstract the notion of a social event (Agarwal et al., 2010), We summarize our experience using a particular kind of event which involves (at least) FrameNet in two rather different projects two people such that at least one of them is aware in natural language processing (NLP). of the other person. If only one person is aware We conclude that NLP can benefit from of the event, we call it Observation (OBS): for FrameNet in different ways, but we sketch example, someone is talking about someone else some problems that need to be overcome. in their absence. If both people are aware of the event, we call it Interaction (INR): for example, 1 Introduction one person is telling the other a story. Our claim We present two projects at Columbia in which we is that links in social networks are in fact made use FrameNet. In these projects, we do not de- up of social events: OBS social events give rise velop basic NLP tools for FrameNet, and we do to one-way links, and INR social events to two- not develop FramNets for new languages: we sim- way links. For more information, see (Agarwal ply use FrameNet or a FrameNet parser in an NLP and Rambow, 2010; Agarwal et al., 2013a; Agar- application.
    [Show full text]
  • Event-Based Knowledge Reconciliation Using Frame Embeddings And
    Knowledge-Based Systems 135 (2017) 192–203 Contents lists available at ScienceDirect Knowle dge-Base d Systems journal homepage: www.elsevier.com/locate/knosys Event-base d knowle dge reconciliation using frame emb e ddings and frame similarity ∗ Mehwish Alam a, Diego Reforgiato Recupero b,d, , Misael Mongiovi c, Aldo Gangemi a,d, Petar Ristoski e a Université Paris 13, 99 avenue JB Clément, Villetaneuse, Paris 93430, France b Università degli Studi di Cagliari, Department of Mathematics and Computer Science, Via Ospedale 72, Cagliari 09124, Italy c CNR, ISTC, Catania, Italy d CNR, ISTC, Via S. Martino della Battaglia 44, Rome, Italy e University of Mannheim, Mannheim, Germany a r t i c l e i n f o a b s t r a c t Article history: This paper proposes an evolution over MERGILO, a tool for reconciling knowledge graphs extracted from Received 6 April 2017 text, using graph alignment and word similarity. The reconciled knowledge graphs are typically used Revised 9 August 2017 for multi-document summarization, or to detect knowledge evolution across document series. The main Accepted 14 August 2017 point of improvement focuses on event reconciliation i.e., reconciling knowledge graphs generated by text Available online 16 August 2017 about two similar events described differently. In order to gather a complete semantic representation of Keywords: events, we use FRED semantic web machine reader, jointly with Framester, a linguistic linked data hub Knowledge reconciliation represented using a novel formal semantics for frames. Framester is used to enhance the extracted event Frame semantics knowledge with semantic frames.
    [Show full text]
  • N-Gram-Based Machine Translation
    N-gram-based Machine Translation ∗ JoseB.Mari´ no˜ ∗ Rafael E. Banchs ∗ Josep M. Crego ∗ Adria` de Gispert ∗ Patrik Lambert ∗ Jose´ A. R. Fonollosa ∗ Marta R. Costa-jussa` Universitat Politecnica` de Catalunya This article describes in detail an n-gram approach to statistical machine translation. This ap- proach consists of a log-linear combination of a translation model based on n-grams of bilingual units, which are referred to as tuples, along with four specific feature functions. Translation performance, which happens to be in the state of the art, is demonstrated with Spanish-to-English and English-to-Spanish translations of the European Parliament Plenary Sessions (EPPS). 1. Introduction The beginnings of statistical machine translation (SMT) can be traced back to the early fifties, closely related to the ideas from which information theory arose (Shannon and Weaver 1949) and inspired by works on cryptography (Shannon 1949, 1951) during World War II. According to this view, machine translation was conceived as the problem of finding a sentence by decoding a given “encrypted” version of it (Weaver 1955). Although the idea seemed very feasible, enthusiasm faded shortly afterward because of the computational limitations of the time (Hutchins 1986). Finally, during the nineties, two factors made it possible for SMT to become an actual and practical technology: first, significant increment in both the computational power and storage capacity of computers, and second, the availability of large volumes of bilingual data. The first SMT systems were developed in the early nineties (Brown et al. 1990, 1993). These systems were based on the so-called noisy channel approach, which models the probability of a target language sentence T given a source language sentence S as the product of a translation-model probability p(S|T), which accounts for adequacy of trans- lation contents, times a target language probability p(T), which accounts for fluency of target constructions.
    [Show full text]
  • Out-Of-Domain Framenet Semantic Role Labeling
    Out-of-domain FrameNet Semantic Role Labeling Silvana Hartmann§†, Ilia Kuznetsov†, Teresa Martin§†, Iryna Gurevych§† Research Training Group AIPHES Ubiquitous§ Knowledge Processing (UKP) Lab Department† of Computer Science, Technische Universitat¨ Darmstadt http://www.ukp.tu-darmstadt.de Abstract on a range of benchmark datasets. This is crucial as the demand for semantic textual analysis of large- Domain dependence of NLP systems is one scale web data keeps growing. of the major obstacles to their application in large-scale text analysis, also restrict- Based on FrameNet (Fillmore et al., 2003), ing the applicability of FrameNet semantic FrameNet SRL extracts frame-semantic structures role labeling (SRL) systems. Yet, current on the sentence level that describe a specific FrameNet SRL systems are still only eval- situation centered around a semantic predicate, uated on a single in-domain test set. For often a verb, and its participants, typically the first time, we study the domain depen- syntactic arguments or adjuncts of the predicate. frame dence of FrameNet SRL on a wide range of The predicate is assigned a label, essentially benchmark sets. We create a novel test set a word sense label, that defines the situation and semantic roles for FrameNet SRL based on user-generated determines the of the participants. web text and find that the major bottleneck The following sentence from FrameNet provides Grinding for out-of-domain FrameNet SRL is the an example of the frame and its roles: frame identification step. To address this [The mill]Grinding cause grindsGrinding [the problem, we develop a simple, yet efficient malt]P atient [to grist]Result.
    [Show full text]