Style Example-Guided Text Generation Using Generative Adversarial Transformers

Style Example-Guided Text Generation Using Generative Adversarial Transformers

Style Example-Guided Text Generation using Generative Adversarial Transformers STYLE EXAMPLE-GUIDED TEXT GENERATION USING GENERATIVE ADVERSARIAL TRANSFORMERS Kuo-Hao Zeng∗, Mohammad Shoeybi & Ming-Yu Liu [email protected], fmshoeybi, [email protected] NVIDIA, Santa Clara, California ABSTRACT We introduce a language generative model framework for generating a styled para- graph based on a context sentence and a style reference example. The framework consists of a style encoder and a texts decoder. The style encoder extracts a style code from the reference example, and the text decoder generates texts based on the style code and the context. We propose a novel objective function to train our framework. We also investigate different network design choices. We conduct extensive experimental validation with comparison to strong baselines to validate the effectiveness of the proposed framework using a newly collected dataset with diverse text styles. Both code and dataset will be released upon publication. 1 INTRODUCTION Text generation refers to the task of generating a new text based on some user input. The task exists in many forms, but arguably the most common form concerns generating a coherent and consistent text based on an input context such as the first few sentences of the target output. This is often achieved by giving the context to a generative language model. Generative language models play a central role in machine learning and natural language processing (NLP). Not only they serve as the main mean for unsupervised feature representation learning but also find use in various applications, including question answering, dialogue agents, summarization, and content creation systems. Thanks to the introduction of novel deep learning architectures and the availability of large-scale training corpora, the state-of-the-art text generation has advanced significantly in recent years. We can now train language models capable of generating fluent and coherent texts that people cannot tell them apart from those written by humans. However, despite the great achievement, existing generative models are limited and inflexible in the sense that a trained model is only capable in generating texts of one style. It can not be used to generate texts of different styles. For instance, a news generative model can only be used to generate news, and a lyric generative model can only be used to generate lyrics. In contrast, humans can compose texts in various styles. arXiv:2003.00674v1 [cs.CL] 2 Mar 2020 To bridge the gap, we propose a style example-guided text generation framework that can generate styled texts based on the style of the example reference text. In our framework, the generator takes two inputs where one is the context input while the other is the style reference example. We use the style reference example to change the generation behavior of our generative model dynamically. For a fixed context, when the provided style reference is a paragraph sampled from a news article, it becomes a news generator. When the provided style reference is a review, it becomes a review generator. In other words, the same generator can generate texts of different styles based on the examples. In Tab. 1, we show example outputs of the proposed framework where we generate texts of dramatically different styles for the same input sentence. The proposed style example-guided text generation framework is based on the generative adversarial networks (GANs), and we utilize the transformer in both the generator and discriminator design. We collect a large dataset containing documents with many different styles for training. Using a novel learning objective function, our network learns to generate styled texts based on the input style example in an unsupervised manner. We conduct extensive experimental validations with comparisons to strong baselines. We also investigate different ways of designing the generator and ∗The work was done during internship at NVIDIA. 1 Style Example-Guided Text Generation using Generative Adversarial Transformers Context – Wiki Style (Machine Learning Intro.) Machine learning (ML) is the scientific study of algorithms and statistical models that computer systems use to perform a specific task without using explicit instructions, relying on patterns and inference instead. It is seen as a subset of artificial intelligence. Machine learning algorithms build a mathematical model based on sample data, known as ”training data”, in order to make predictions or decisions without being explicitly programmed to perform the task.[1][2]: Reference – News Style Reference – Yelp Reviews Style (CNN)Remember that robot dog that went viral a few years ago? The one that can Yumm run uphill, climb stairs, and looks like the killer robot dogs in the TV show ”Black Mirror”? It’s now available for select customers. The line was AWFUL! Came by Saturday around 5:20pm and the line was CRAZY. I do feel like they’re understaffed. Despite that, it was my first time and i got the chance Boston Dynamics, the company behind the dog, which is named Spot, released a to taste test two flavors, matcha and milk tea! Both were great!! video on Tuesday advertising its launch. It shows Spot navigating challenging terrain, picking up construction objects, moving through bad weather, and picking itself up I got the : after a fall. The video and the product launch page also show off Spot’s capabilities – - Matcha swirl with a custard fish, sprinkles and coconut flakes : The custard was so for instance, it can run at 1.6 meters per second (about 5.2 feet per second), has 360- good! Paired great with the ice cream. degree cameras that give it a full range of vision, is dust- and water-proof, and can operate in temperatures ranging from -20 to 45 degrees Celsius (-4 to 113 Fahrenheit). Staff were so friendly! Especially the cashier (guy) and person who swirled the ice cream that shift! She was super patient as I, myself is very indecisive, haha! Most definitely coming by again!! Generated Generated Machine Learning is an ancient concept that has entered the mainstream today, which Since the day machines started boosting up, dozens of companies have recently done means that it can survive in the face of such heightened competition. In its application, so. In fact, I highly recommend RoboSports, an example of small-batch ice cream Flurry wants to improve the efficiency of AI driven computing machines, and with a places that have taken up artificial advances in AI; their ice cream is the perfect size, large footprint. As is seen on me, Flurry has designed a large extension that (unfortu- slim and hot, it was great. If you are a connoisseur of ice cream, check out this nately) can be used in its development, but with lower speed lag. Now that the hybrid wonderful ice cream shop. technology is here, it’s there for the taking. There may also be a Something About Mary in the future. ”So it’s nice to be able to mount a cat across your desk without requiring any backup access.” Table 1: Given the same context sentences but different style reference examples, our generation model can generate paragraphs of different styles. The output paragraph shares the same style as the reference example. compare their performance. Through detailed quantitative and user study results, we prove the effectiveness of the proposed framework for the style example-guided text generation task. 2 RELATED WORK Language modeling has seen many advancements in recent years, which has resulted in significant improvements on various NLP tasks. Early language models focused on using n-grams to represent a text distribution. Bengio et al. (2003) introduced a neural language model in a shift from more tra- ditional n-gram models. Many works later (Mikolov et al. (2013); Pennington et al. (2014)) focused on word embeddings as a way to represent tokens within the text. More recently, Peters et al. (2018) used bi-directional LSTMs to obtain deep contextualized word representation. However, RNNs can only represent a limited context. Vaswani et al. (2017) introduced the transformer networks which use the connections between long-distance word pairs embedded in attention mechanisms and can easily enable the learning of long-term dependency. Many later models (Devlin et al. (2019); Liu et al. (2019d); Dai et al. (2019); Yang et al. (2019)) used transformer model and obtained signifi- cant improvements on downstream tasks (Wang et al. (2019); Rajpurkar et al. (2016); Zellers et al. (2018)). Lately, (Radford et al. (2019)) introduced GPT-2, a generative left-to-right language model based on the transformer and showed that these models are able to generate coherent text when pre- trained on a large corpus. Shoeybi et al. (2019) further scaled up the GPT-2 model and demonstrated improved performance. Our work differs from the prior works because we aim for allowing user flexible control over the style of the generated text. Texts generation includes review generation (Radford et al. (2018); Zang & Wan (2017)), sentiment texts generation (Wang & Wan (2018); Hu et al. (2017); Merity et al. (2017)), Wikipedia generation (Liu et al. (2018); Lebret et al. (2016)), fake news generation (Bakhtin et al. (2019); Zellers et al. (2019)), abstractive summarization (Li et al. (2018); Zhang et al. (2019); Pasunuru et al. (2017)), and conversation/dialogue system (Vinyals & Le (2015); Budzianowski & Vulic´ (2019)). Although many of them trained a transformer on large-scale corpora, their results were limited in their specific domain (e.g., reviews, news, etc.) because they either utilized domain-specific priors in their model design or were not designed to generate texts in many different domains or styles. Control on texts generation. In addition, there are literature utilizing insertion-base (Stern et al. (2019); Chan et al. (2019)), GAN-based (Yu et al. (2017); d’Autume et al.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    20 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us