A Reinforced Topic-Aware Convolutional Sequence-To-Sequence Model for Abstractive Text Summarization

A Reinforced Topic-Aware Convolutional Sequence-To-Sequence Model for Abstractive Text Summarization

Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18) A Reinforced Topic-Aware Convolutional Sequence-to-Sequence Model for Abstractive Text Summarization Li Wang1, Junlin Yao2, Yunzhe Tao3, Li Zhong1, Wei Liu4, Qiang Du3 1 Tencent Data Center of SNG 2 ETH Zurich¨ 3 Columbia University 4 Tencent AI Lab [email protected], [email protected], [email protected], [email protected], [email protected], [email protected] Abstract There are two genres of automatic summarization tech- niques, namely, extraction and abstraction. The goal of ex- In this paper, we propose a deep learning approach tractive summarization [Neto et al., 2002] is to produce a to tackle the automatic summarization tasks by summary by selecting important pieces of the source doc- incorporating topic information into the convolu- ument and concatenating them verbatim, while abstractive tional sequence-to-sequence (ConvS2S) model and summarization [Chopra et al., 2016] generates summaries using self-critical sequence training (SCST) for op- based on the core ideas of the document, therefore the sum- timization. Through jointly attending to topics and maries could be paraphrased in more general terms. Other word-level alignment, our approach can improve than extraction, abstractive methods should be able to prop- coherence, diversity, and informativeness of gen- erly rewrite the core ideas of the source document and assure erated summaries via a biased probability genera- that the generated summaries are grammatically correct and tion mechanism. On the other hand, reinforcement human readable, which is close to the way how humans do training, like SCST, directly optimizes the pro- summarization and thus is of interest to us in this paper. posed model with respect to the non-differentiable metric ROUGE, which also avoids the exposure Recently, deep neural network models have been widely bias during inference. We carry out the experimen- used for NLP tasks. In particular, the attention based tal evaluation with state-of-the-art methods over the sequence-to-sequence framework [Bahdanau et al., 2014] Gigaword, DUC-2004, and LCSTS datasets. The with recurrent neural networks (RNNs) [Sutskever et al., empirical results demonstrate the superiority of our 2014] prevails in the NLP tasks. However, RNN-based mod- proposed method in the abstractive summarization. els are more prone to gradient vanishing due to their chain structure of non-linearities compared to the hierarchical struc- ture of CNN-based models [Dauphin et al., 2016]. In ad- 1 Introduction dition, the temporal dependence among the hidden states Automatic text summarization has played an important role in of RNNs prevents parallelization over the elements of a se- a variety of natural language processing (NLP) applications, quence, which makes the training inefficient. In this pa- such as news headlines generation [Kraaij et al., 2002] and per, we propose a new approach based on the convolutional feeds stream digests [Barzilay and McKeown, 2005]. It is sequence-to-sequence (ConvS2S) framework [Gehring et al., of interest to generate informative and representative natural 2017] jointly with a topic-aware attention mechanism. To language summaries which are capable of retaining the main the best of our knowledge, this is the first work for auto- ideas of source articles. The key challenges in automatic text matic abstractive summarization that incorporates the topic summarization are correctly evaluating and selecting impor- information, which can provide themed and contextual align- tant information, efficiently filtering redundant contents, and ment information into deep learning architectures. In addi- properly aggregating related segments and making human- tion, we also optimize our proposed model by employing the readable summaries. Compared to other NLP tasks, the au- reinforcement training [Paulus et al., 2017]. The main contri- tomatic summarization has its own difficulties. For example, butions of this paper include: unlike machine translation tasks where input and output se- quences often share similar lengths, summarization tasks are • We propose a joint attention and biased probability gen- more likely to have input and output sequences greatly im- eration mechanism to incorporate the topic information balanced. Besides, machine translation tasks usually have into an automatic summarization model, which intro- some direct word-level alignment between input and output duces contextual information to help the model generate sequences, which is less obvious in summarization. more coherent summaries with increased diversity. 4453 Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18) • We employ the self-critical sequence training technique in ConvS2S to directly optimize the model with respect to the non-differentiable summarization metric ROUGE, which also remedies the exposure bias issue. • Extensive experimental results on three datasets demon- strate that by fully exploiting the power of the ConvS2S architecture enhanced by topic embedding and SCST, our proposed model yields high accuracy for abstractive summarization, advancing the state-of-the-art methods. 2 Related Work Automatic text summarization has been widely investigated. Various methods [Neto et al., 2002] focus on the extractive summarization, which select important contents of text and combine them verbatim to produce a summary. On the other hand, abstractive summarization models are able to produce a grammatical summary with a novel expression, most of which [Rush et al., 2015; Chopra et al., 2016; Nallapati et al., 2016a] are built upon the neural attention-based sequence-to- sequence framework [Sutskever et al., 2014]. The predominant models are based on RNNs [Nallapati et al., 2016b; Shen et al., 2016; Paulus et al., 2017], where the Figure 1: A graphical illustration of the topic-aware convolutional encoder and decoder are constructed using either Long Short- architecture. Word and topic embeddings of the source sequence are Term Memory (LSTM) [Hochreiter and Schmidhuber, 1997] encoded by the associated convolutional blocks (bottom left and bot- or Gated Recurrent Unit (GRU) [Cho et al., 2014]. However, tom right). Then we jointly attend to words and topics by comput- ing dot products of decoder representations (top left) and word/topic very few methods have explored the performance of convolu- encoder representations. Finally, we produce the target sequence tional structure on summarization tasks. Compared to RNNs, through a biased probability generation mechanism. convolutional neural networks (CNNs) enjoy several advan- tages, including efficient training by leveraging parallel com- puting, and mitigating the gradient vanishing problem due to structure, and a reinforcement learning procedure. The graph- fewer non-linearities [Dauphin et al., 2016]. Notably, the re- ical illustration of the topic-aware convolutional architecture cently proposed gated convolutional network [Dauphin et al., can be found in Figure 1. 2016; Gehring et al., 2017] outperforms RNN-based models in the language modeling and machine translation tasks. 3.1 ConvS2S Architecture While the ConvS2S model is also evaluated on the ab- We exploit the ConvS2S architecture [Gehring et al., 2017] stractive summarization [Gehring et al., 2017], there are sev- as the basic infrastructure of our model. In this paper, two eral limitations. First, the model is trained by minimizing convolutional blocks are employed, associated with the word- a maximum-likelihood loss which is sometimes inconsistent level and topic-level embeddings, respectively. We introduce with the metric that is evaluated on the sentence level, e.g., the former in this section and the latter in next, along with the ROUGE [Lin, 2004]. In addition, the exposure bias [Ran- new joint attention and the biased generation mechanism. zato et al., 2015] occurs due to only exposing the model to the training data distribution instead of its own predictions. Position Embeddings More importantly, the ConvS2S model utilizes only word- Let x = (x1; : : : ; xm) denote the input sentence. We first level alignment which may be insufficient for summariza- embed the input elements (words) in a distributional space Rd tion and prone to incoherent generated summaries. There- as w = (w1; : : : ; wm), where wi 2 are rows of a RV ×d fore, the higher level alignment could be a potential assist. randomly initialized matrix Dword 2 with V being For example, the topic information has been introduced to a the size of vocabulary. We also add a positional embed- Rd RNN-based sequence-to-sequence model [Xing et al., 2017] ding, p = (p1; : : : ; pm) with pi 2 , to retain the or- for chatbots to generate more informative responses. der information. Thus, the final embedding for the input is e = (w1 +p1; : : : ; wm +pm). Similarly, let q = (q1; : : : ; qn) denote the embedding for output elements that were already 3 Reinforced Topic-Aware Convolutional generated by the decoder and being fed back to the next step. Sequence-to-Sequence Model Convolutional Layer In this section, we propose the Reinforced Topic-Aware Con- Both encoder and decoder networks are built by stacking sev- volutional Sequence-to-Sequence model, which consists of a eral convolutional layers. Suppose that the kernel has width convolutional architecture with both input words and topics, of k and the input embedding dimension is d. The convolu- a joint multi-step

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    8 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us