Controllable Text Simplification with Explicit Paraphrasing

Controllable Text Simplification with Explicit Paraphrasing

Controllable Text Simplification with Explicit Paraphrasing Mounica Maddela1, Fernando Alva-Manchego2, Wei Xu1 1School of Interactive Computing, Georgia Institute of Technology 2Department of Computer Science, University of Sheffield {mounica.maddela, wei.xu}@cc.gatech.edu [email protected] Abstract OLen %new %eq %split Complex (input) 20.7 0.0 100.0 0.0 Text Simplification improves the readability Narayan and Gardent(2014) y 10.4 0.7 0.8 0.4 Zhang and Lapata(2017) y 13.8 8.1 16.8 0.0 of sentences through several rewriting trans- Dong et al.(2019) y 10.9 8.4 4.6 0.0 formations, such as lexical paraphrasing, dele- Kriz et al.(2019) y 10.8 11.2 1.2 0.0 tion, and splitting. Current simplification sys- LSTM 17.0 6.1 28.4 1.2 tems are predominantly sequence-to-sequence Our Model 17.1 17.0 3.0 31.8 models that are trained end-to-end to perform Simple (reference) 17.9 29.0 0.0 30.0 all these operations simultaneously. However, such systems limit themselves to mostly delet- Table 1: Output statistics of 500 random sentences ing words and cannot easily adapt to the re- from the Newsela test set. Existing systems rely on quirements of different target audiences. In deletion and do not paraphrase well. OLen, %new, this paper, we propose a novel hybrid approach %eq and %split denote the average output length, per- that leverages linguistically-motivated rules centage of new words added, percentage of system out- for splitting and deletion, and couples them puts that are identical to the inputs, and percentage of with a neural paraphrasing model to produce sentence splits, respectively. yWe used the system out- varied rewriting styles. We introduce a new puts shared by their authors. data augmentation method to improve the para- phrasing capability of our model. Through au- models trained end-to-end, which have greatly in- tomatic and manual evaluations, we show that creased the fluency of the outputs (Zhang and Lap- our proposed model establishes a new state-of- ata, 2017; Nisioi et al., 2017; Zhao et al., 2018; Kriz the art for the task, paraphrasing more often et al., 2019; Dong et al., 2019; Jiang et al., 2020). than the existing systems, and can control the However, these systems mostly rely on deletion degree of each simplification operation applied and tend to generate very short outputs at the cost 1 to the input texts. of meaning preservation (Alva-Manchego et al., 1 Introduction 2017). Table1 shows that they neither split sen- tences nor paraphrase well as reflected by the low Text Simplification aims to improve the readability percentage of splits (< 1%) and new words intro- of texts with simpler grammar and word choices duced (< 11.2%). While deleting words is a viable while preserving meaning (Saggion, 2017). It pro- (and the simplest) way to reduce the complexity of vides reading assistance to children (Kajiwara et al., sentences, it is suboptimal and unsatisfying. Pro- 2013), non-native speakers (Petersen and Osten- fessional editors are known to use a sophisticated dorf, 2007; Pellow and Eskenazi, 2014; Paetzold, combination of deletion, paraphrasing, and sen- 2016), and people with reading disabilities (Rello tence splitting to simplify texts (Xu et al., 2015). et al., 2013). It also helps with downstream nat- Another drawback of these end-to-end neural ural language processing tasks, such as parsing systems is the lack of controllability. Simplification (Chandrasekar et al., 1996), semantic role labelling is highly audience dependant, and what constitutes (Vickrey and Koller, 2008), information extraction simplified text for one group of users may not be (Miwa et al., 2010), and machine translation (MT, acceptable for other groups (Xu et al., 2015; Lee Chen et al., 2012; Štajner and Popovic, 2016). and Yeung, 2018). An ideal simplification system Since 2016, nearly all text simplification sys- should be able to generate text with varied char- tems have been sequence-to-sequence (seq2seq) acteristics, such as different lengths, readability 1Our code and data are available at https://github. levels, and number of split sentences, which can be com/mounicam/controllable_simplification. difficult to control in end-to-end systems. 3536 Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 3536–3553 June 6–11, 2021. ©2021 Association for Computational Linguistics Figure 1: Overview of our proposed model for text simplification, which can perform a controlled combination of sentence splitting, deletion, and paraphrasing. To address these issues, we propose a novel 2 Our Approach hybrid approach that combines linguistically- motivated syntactic rules with data-driven neural Figure1 shows an overview of our hybrid approach. models to improve the diversity and controllabil- We combine linguistic rules with data-driven neural ity of the simplifications. We hypothesize that the models to improve the controllability and diversity seq2seq generation model will learn lexical and of the outputs. Given an input complex sentence structural paraphrases more efficiently from the x, we first generate a set of intermediate simplifica- parallel corpus, when we offload some of the bur- tions V = fv1; v2;:::; vng that have undergone den of sentence splitting (e.g., split at comma) and splitting and deletion (§2.1). These intermediate deletion (e.g., remove trailing preposition phrases) sentences are then used for two purposes: (1) Se- decisions to a separate component. Previous hy- lected by a pairwise neural ranking model (§2.2) brid approaches for simplification (Narayan and based on the simplification quality and then rewrit- Gardent, 2014; Siddharthan and Mandya, 2014; ten by the paraphrasing component; (2) Used for Sulem et al., 2018c) used splitting and deletion data augmentation to improve the diversity of the rules in a deterministic step before applying an paraphrasing model (§2.3). MT-based paraphrasing model. In contrast, our 2.1 Splitting and Deletion approach provides a more flexible and dynamic in- tegration of linguistic rules with the neural models We leverage the state-of-the-art system for struc- through ranking and data augmentation (Figure1). tural simplification, called DisSim (Niklaus et al., We compare our method to several state-of-the- 2019), to generate candidate simplifications that fo- 2 art systems in both automatic and human evalua- cus on splitting and deletion. The English version tions. Our model achieves overall better perfor- of DisSim applies 35 hand-crafted grammar rules mance measured by SARI (Xu et al., 2016) and to break down a complex sentence into a set of hi- other metrics, showing that the generated outputs erarchically organized sub-sentences (see Figure1 are more similar to those written by human editors. for an example). We choose a rule-based approach We also demonstrate that our model can control the for sentence splitting because it works really well. extent of each simplification operation by: (1) im- In our pilot experiments, DisSim successfully split posing a soft constraint on the percentage of words 92% of 100 complex sentences from the training to be copied from the input in the seq2seq model, data with more than 20 words, and introduced er- thus limiting lexical paraphrasing; and (2) select- rors for only 6.8% of these splits. We consider ing candidates that underwent a desired amount these sub-sentences as candidate simplifications of splitting and/or deletion. Finally, we create a for the later steps, except those that are extremely new test dataset with multiple human references short or long (compression ratio 2= [0.5, 1.5]). The for Newsela (Xu et al., 2015), the widely used text compression ratio is calculated as the number of simplification corpus, to specifically evaluate lexi- 2https://github.com/Lambda-3/ cal paraphrasing. DiscourseSimplification 3537 words in a candidate simplification vi (which may candidate pair, and learn to minimize the pairwise contain one or more sub-sentences) divided by that ranking violations using hinge loss: of the original sentence x. To further increase the variety of generated can- m nk nk 1 X 1 X X k k didates, we supplement DisSim with a Neural LMR = max(0; 1 − l d ) m n2 ij ij Deletion and Split module trained on the text sim- k=1 k i=1 j=1;i6=j k k k plification corpus (§3.1). We use a Transformer dij = g(vi ) − g(vj ) seq2seq model with the same configuration as the k ∗ k k ∗ k k base model for paraphrasing (§2.3). Given the in- lij = sign g (vi ; y ) − g (vj ; y ) (2) put sentence x, we constrain the beam search to generate 10 outputs with splitting and another 10 where g(:) is a feedforward neural network, m is outputs without splitting. Then, we select the out- the number of training complex-simple sentence puts that do not deviate substantially from x (i.e., pairs, k is the index of training examples, and Jaccard similarity > 0.5). We add outputs from the nk represents the number of generated candidates two systems to the candidate pool V . (§2.1). On average, nk is about 14.5 for a sentence of 30 words, and can be larger for longer sentences. 2.2 Candidate Ranking We consider 10 randomly sampled candidates for each complex sentence during training. We design a neural ranking model to score all the candidates that underwent splitting and deletion, Features. For the feedforward network g(:), we V = fv1; v2;:::; vng, then feed the top-ranked use the following features: number of words in one to the lexical paraphrasing model for the final vi and x, compression ratio of vi with respect to output. We train the model on a standard text sim- x, Jaccard similarity between vi and x, the rules plification corpus consisting of pairs of complex applied on x to obtain vi, and the number of rule sentence x and manually simplified reference y.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    18 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us