An Encoder-Decoder Approach for Cross-Lingual Semantic Role Labeling

An Encoder-Decoder Approach for Cross-Lingual Semantic Role Labeling

Translate and Label! An Encoder-Decoder Approach for Cross-lingual Semantic Role Labeling Angel Daza and Anette Frank Leibniz ScienceCampus “Empirical Linguistics and Computational Language Modeling” Department of Computational Linguistics Heidelberg University 69120 Heidelberg, Germany fdaza,[email protected] Abstract We propose a Cross-lingual Encoder-Decoder model that simultaneously translates and gen- erates sentences with Semantic Role Labeling annotations in a resource-poor target language. Unlike annotation projection techniques, our model does not need parallel data during in- ference time. Our approach can be applied in monolingual, multilingual and cross-lingual settings and is able to produce dependency- Figure 1: We propose an Encoder-Decoder model that based and span-based SRL annotations. We translates a sentence into a target language and applies benchmark the labeling performance of our SRL labeling to the translated words. In this example model in different monolingual and multilin- we translate from English to German and label roles for gual settings using well-known SRL datasets. the predicate have. We then train our model in a cross-lingual set- ting to generate new SRL labeled data. Finally, we measure the effectiveness of our method by predicate (Marcheggiani et al., 2017; He et al., using the generated data to augment the train- 2017), and even jointly predict predicates inside ing basis for resource-poor languages and per- sentences (He et al., 2018; Cai et al., 2018). While form manual evaluation to show that it pro- these approaches alleviate the need for pipeline duces high-quality sentences and assigns accu- rate semantic role annotations. Our proposed models, they require sufficient amounts of training architecture offers a flexible method for lever- data to perform adequately. To date, such models aging SRL data in multiple languages. have been tested primarily for English, which of- fers a considerable amount of high-quality train- 1 Introduction ing data compared to other languages. The lack Semantic Role Labeling (SRL) extracts semantic of sufficiently large SRL datasets makes it hard to predicate-argument structure from sentences. This straightforwardly apply the same architectures to has proven to be useful in Neural Machine Trans- other languages and calls for methods to augment lation (NMT) (Marcheggiani et al., 2018), Multi- the training data in lower-resource languages. document-summarization (Khan et al., 2015), There is significant prior work on SRL data aug- AMR parsing (Wang et al., 2015) and Read- mentation (Hartmann et al., 2017), annotation pro- ing Comprehension (Mihaylov and Frank, 2019). jection for monolingual (Furstenau¨ and Lapata, SRL consists of three steps: i) predicate detection, 2012; Hartmann et al., 2016), and cross-lingual ii) argument identification and iii) role classifica- SRL (Pado´ and Lapata, 2009; van der Plas et al., tion. In this work we focus on PropBank SRL 2011; Akbik et al., 2015, 2016). A drawback of (Palmer et al., 2005), which has proven its validity cross-lingual projection is that even at prediction across languages (van der Plas et al., 2010). While time it requires parallel sentences, a semantic role former SRL systems rely on syntactic features labeler on the source side, as well as syntactic (Punyakanok et al., 2008;T ackstr¨ om¨ et al., 2015), information for both language sides. Thus, it is recent neural approaches learn to model both ar- desirable to design an architecture that can make gument detection and role classification given a use of existing annotations in more than one lan- 603 Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, pages 603–615, Hong Kong, China, November 3–7, 2019. c 2019 Association for Computational Linguistics guage and that learns to translate input sentences to another language while transferring semantic role annotations from the source to the target. Techniques for low-resource Neural Machine Translation (NMT) show the positive impact on target predictions by adding more than one language during training, such as Multi-source NMT (Zoph and Knight, 2016) and Multilingual Figure 2: Monolingual Enc-Dec model for SRL with NMT (Johnson et al., 2017; Firat et al., 2016a), copying (Daza and Frank, 2018). We generalize this whereas Mulcaire et al.(2018) show the advan- architecture to multilingual and cross-lingual SRL. tages of training a single polyglot SRL system that improves over monolingual baselines in lower- contain word tokens of different languages if de- resource settings. In this work, we propose a gen- sired. This means that we could train an Enc-Dec eral Encoder-Decoder (Enc-Dec) architecture for model that learns not only to label a sentence, but SRL (see Figure1). We extend our previous Enc- to jointly translate it while applying SRL annota- Dec approach for SRL (Daza and Frank, 2018) to a tions directly to the target language. Moreover, cross-lingual model that translates sentences from following conceptually the multilingual Enc-Dec a source language to a (lower-resource) target lan- model proposed by Johnson et al.(2017), we can guage, and during decoding jointly labels it with train a single model that allows for joint training SRL annotations.1 with multiple language pairs while sharing param- Our contributions are as follows: eters among them. We apply a similar joint multi- • We propose the first cross-lingual multilin- lingual learning method to produce structured out- gual Enc-Dec model for PropBank SRL. put sequences in the form of translations enriched • We show that our cross-lingual model can with SRL annotations on the (lower-resource) tar- generate new labeled sentences in a target get language (cf. Figure3). We will apply this language without the need of explicit syntac- universal structure-inducing Enc-Dec model to the tic or semantic annotations at inference time. Semantic Role Labeling task, and show that it can • Cross-lingual evaluation against a labeled be deployed in three different settings: gold standard achieves good performance, i) monolingual: encode a sentence in a given comparable to monolingual SRL results. language and learn to decode a labeled sequence by reproducing the source words and inserting the • Augmenting the training set of a lower- appropriate structure-indicating labels in the out- resource language with sentences generated put (cf. Figure2). A copying mechanism (Gu by the cross-lingual model achieves im- et al., 2016) allows this model to reproduce the in- proved F1 scores on the benchmark dataset. put sentence as faithfully as possible. • Our universal Enc-Dec model lends itself to ii) one-to-one multilingual: train a single, joint monolingual, multilingual and crosslingual model to generate n different structure-enriched SRL and yields competitive performance. target languages given inputs in the same lan- guage. For example: Labeled English (EN-SRL) 2 An Extensible Model for SRL given an EN sentence or Labeled German (DE- SRL) given a DE sentence. This multilingual 2.1 One Model to Treat Them All model still relies on copying to relate each labeled We define the SRL task as a sequence transduc- output sentence to its corresponding input coun- tion problem: given an input sequence of tokens terpart. However, unlike (i), it has the advantage X = x1; :::; xi, the system is tasked to generate of sharing parameters among languages. a sequence Y = y1; :::; yj consisting of words in- iii) cross-lingual: generate outputs in n differ- terleaved with SRL annotations. Defining the task ent target languages given inputs in m different in this fashion allows X and Y to be of different source languages, for example: Labeled German lengths and therefore target sequences may also (DE-SRL) and Labeled French (FR-SRL) given an 1Code is available at: https://github.com/ EN sentence (see Figure3). In this setting, we Heidelberg-NLP/SRL-S2S. do not restrict the model to copy words from the 604 source sentence but train it to translate them. mechanism proposed by Bahdanau et al.(2015), In Section 2.2 we describe how the basic Enc- which computes a context vector at each time step Dec model for SRL is constructed and in Section t based on H and the current decoder state. 2.3 we describe the additional components that al- Decoder. We use a single-layer Decoder with low us to generalize this architecture to the one-to- LSTM cells (Hochreiter and Schmidhuber, 1997) one multilingual and cross-lingual scenarios. and a copying mechanism. It emits an output to- ken yt from a learned score g over the vocabulary 2.2 Encoder-Decoder Architecture at each time step t given its state st, the previous We reimplement and extend the Enc-Dec model output token yt−1, and the attention context vector with attention (Bahdanau et al., 2015) and copying ct. In addition, a copying score c is calculated. (Gu et al., 2016) mechanisms for SRL proposed by The decoder learns from these scores when to gen- Daza and Frank(2018). This model encodes the erate a new token and when to copy from the en- source sentence and decodes the input sequence coded hidden states H. Formally we compute the of words (in the same language) interleaved with scores as: SRL labels. (y = v ) = W [s ; c ]; v V[L Data Representation. Similar to other prior g t i o t t i T (1) work (Liu et al., 2018) and our own (Daza and c(yt = xj) = σ(hj Wc)st; xj X Frank, 2018), we linearize the SRL structure in where W N×2ds and W dh×ds are learn- order to process it as a sequence of symbols suit- o R c R able parameters and s , c are the current de- able for the Enc-Dec architecture.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    13 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us