The Importance of Generation Order in Language Modeling

Total Page:16

File Type:pdf, Size:1020Kb

The Importance of Generation Order in Language Modeling The Importance of Generation Order in Language Modeling Nicolas Ford∗ Daniel Duckworth Mohammad Norouzi George E. Dahl Google Brain fnicf,duckworthd,mnorouzi,[email protected] Abstract There has been interest in moving beyond the left-to-right generation order by developing alter- Neural language models are a critical compo- native multi-stage strategies such as syntax-aware nent of state-of-the-art systems for machine neural language models (Bowman et al., 2016) translation, summarization, audio transcrip- and latent variable models of text (Wood et al., tion, and other tasks. These language models are almost universally autoregressive in nature, 2011). Before embarking on a long-term research generating sentences one token at a time from program to find better generation strategies that left to right. This paper studies the influence of improve modern neural networks, one needs ev- token generation order on model quality via a idence that the generation strategy can make a novel two-pass language model that produces large difference. This paper presents one way of partially-filled sentence “templates” and then isolating the generation strategy from the general fills in missing tokens. We compare various neural network design problem. Our key techni- strategies for structuring these two passes and cal contribution involves developing a flexible and observe a surprisingly large variation in model quality. We find the most effective strategy tractable architecture that incorporates different generates function words in the first pass fol- generation orders, while enabling exact computa- lowed by content words in the second. We be- tion of the log-probabilities of a sentence. Our ex- lieve these experimental results justify a more periments demonstrate that even when using a few extensive investigation of generation order for simple two-pass generation orders, the differences neural language models. between good and bad orderings are substantial. We consider ways of reordering the tokens 1 Introduction within a sequence based on their identities. The Neural networks have been extremely successful best ordering we tried generates function words statistical models of text in language modeling and first and content words last, which cuts against the machine translation. Despite differences in model idea of committing to the general topic of a sen- architectures, state of the art neural nets gener- tence first and only then deciding exactly how to ate sequences from left to right (Vaswani et al., phrase it. We offer some possible explanations in 2017; Jozefowicz et al., 2016; Wu et al., 2016). Section3, and we conclude that our experimen- Although in some sense humans produce and con- tal results justify a more extensive investigation of sume language from left to right as well, there are the generation order for language and translation many other intuitively appealing ways to gener- models. ate text. For instance, language is slow enough 2 Two-pass Language Models on a neurological time scale for multiple passes of generation that incorporate feedback to occur. We develop a family of two-pass language mod- Linguistic intuition might suggest that we should els that depend on a partitioning of the vocabu- first generate some abstract representation of what lary into a set of first-pass and second-pass tokens we want to say and then serialize it, a process that to generate sentences. We perform a preprocess- seems more universally appropriate given the ex- ing step on each sequence y, creating two new se- istence of languages with freer word order such as quences y(1) and y(2). The sequence y(1), which Czech and Polish. we call the template, has the same length as y, ∗Work done as a member of the Google AI Residency and consists of the first-pass tokens from y to- program (g.co/airesidency) gether with a special placeholder token wherever 2942 Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 2942–2946 Brussels, Belgium, October 31 - November 4, 2018. c 2018 Association for Computational Linguistics sentence common first rare first function first content first odd first ” all you need to do ” all you to if you need do ” all you to if you need do ” all you need if you want the na- the ’s on want nation the ’s on your want nation press you the nation ’s tion ’s press camped is to you had a press camped your is to you a camped doorstep press camped on your on your doorstep is to [UNK] in , ” he doorstep say in , ” he in his say once had doorstep say you say you once had a in his . [EOS] once 1947 . [EOS] [UNK] 1947 once had [UNK] in 1947 , ” noted memorably noted memorably ” noted his . he noted memorably in diary [EOS] diary [EOS] [EOS] his diary . [EOS] the team announced the that the , team announced the that the team announced the team announced thursday that the 6- [UNK] will in thursday 6-foot-1 , will in thursday 6-foot-1 the 6-foot-1 foot-1 , [UNK] starter the . [EOS] starter remain through the . [UNK] starter will remain will remain in detroit detroit through [EOS] remain detroit through the 2013 . through the 2013 sea- 2013 season [EOS] 2013 season [EOS] [EOS] son . [EOS] scotland ’s next game ’s is a the scotland next game ’s is a against scotland next game ’s next game is a friendly against at on . [EOS] friendly against the at on . friendly the czech republic at the czech republic at czech republic ham- [EOS] czech republic ham- hampden on 3 march . hampden on 3 march . pden 3 march pden 3 march [EOS] [EOS] [EOS] [EOS] of course , millions of of , of course millions of , of a course millions of of additional additional homeown- a : they of additional homeown- : they of ” additional home- big ers did make a big mis- ” ” and [UNK] ers did make big ” and to owners did make they advantage of take : they took ad- to they ’t . mistake took ad- they . [EOS] big mistake ” liar ” and other vantage of ” liar loans [EOS] vantage liar loans took advantage deals buy homes ” and other [UNK] other deals liar loans other they couldn afford . deals to buy homes buy homes couldn [UNK] deals buy [EOS] they couldn ’t afford . afford [EOS] homes couldn ’t [EOS] afford [EOS] Table 1: Some example sentences from the dataset and their corresponding templates. The placeholder token is indicated by “ ”. y had a second-pass token. The sequence y(2) has sequence model that translates the template into length equal to the number of these placeholders, the complete sentence. There are three places in and consists of the second-pass tokens from y in this model where word embeddings appear — the order. first-phase decoder, the second-phase encoder, and We use a neural language model p1 to generate the second-phase decoder — and all three sets (1) y , and then a conditional translation model p2 of parameters are shared. The output layer also to generate y(2) given y(1). Note that, since the shares the embedding parameters.1 division of the vocabulary into first- and second- For the second pass, we include the entire target pass tokens is decided in advance, there is a one- sentence, not just the second-pass tokens, on the to-one correspondence between sequences y and output side. In this way, when generating a token, pairs (y(1); y(2)). The total probability of y is then the decoder is allowed to examine all tokens to the left of its position. However, only the second-pass (1) (2) (1) p(y) = p1(y ) p2(y j y ) : (1) tokens count toward the loss, since in the other po- sitions the correct token is already known. Our Two-pass language models present a unique op- loss function is then the sum of all of these num- portunity to study the importance of generation or- bers (from both copies) divided by the length of der because, since the template is a deterministic the original sentence, which is the log-perplexity function of y, the probability of y can be com- that our model assigns to the sentence. puted exactly. This is in contrast to a language We tried five different ways of splitting the vo- model using a latent generation order, which re- cabulary: quires a prohibitive marginalization over permu- Common First and Rare First: The vocabu- tations to compute the exact probabilities. Given lary was sorted by frequency and then a cutoff was the tractable nature of the model, exact learning chosen, splitting the vocabulary into “common” based on log-likelihood is possible, and we can and “rare” tokens. The location of the cutoff2 was compare different vocabulary partitioning strate- chosen so that the number of common tokens and gies both against each other and against a single- the number of rare tokens in the average sentence pass language model. were approximately the same. In “common first” Our implementation consists of two copies of 1 the Transformer model from Vaswani et al.(2017). This behavior is enabled in the publicly available im- plementation of Transformer using the hyperparameter called The first copy just generates the template, so it has shared embedding and softmax weights. no encoder. The second copy is a sequence-to- 2In our experiments on LM1B, this is at index 78. 2943 we place the common tokens in the first pass, and size of 2048, and 8 attention heads, except that the in “rare first” we start with the rare tokens. enhanced baseline used a hidden size of 704. We Function First and Content First: We parsed used a batch size of 4096. All models were trained about 1% of LM1B’s training set using Parsey Mc- using ADAM (Kingma and Ba, 2014), with β1 = −6 Parseface (Andor et al., 2016) and assigned each 0:85, β2 = 0:997, and = 10 .
Recommended publications
  • Artificial Intelligence in Health Care: the Hope, the Hype, the Promise, the Peril
    Artificial Intelligence in Health Care: The Hope, the Hype, the Promise, the Peril Michael Matheny, Sonoo Thadaney Israni, Mahnoor Ahmed, and Danielle Whicher, Editors WASHINGTON, DC NAM.EDU PREPUBLICATION COPY - Uncorrected Proofs NATIONAL ACADEMY OF MEDICINE • 500 Fifth Street, NW • WASHINGTON, DC 20001 NOTICE: This publication has undergone peer review according to procedures established by the National Academy of Medicine (NAM). Publication by the NAM worthy of public attention, but does not constitute endorsement of conclusions and recommendationssignifies that it is the by productthe NAM. of The a carefully views presented considered in processthis publication and is a contributionare those of individual contributors and do not represent formal consensus positions of the authors’ organizations; the NAM; or the National Academies of Sciences, Engineering, and Medicine. Library of Congress Cataloging-in-Publication Data to Come Copyright 2019 by the National Academy of Sciences. All rights reserved. Printed in the United States of America. Suggested citation: Matheny, M., S. Thadaney Israni, M. Ahmed, and D. Whicher, Editors. 2019. Artificial Intelligence in Health Care: The Hope, the Hype, the Promise, the Peril. NAM Special Publication. Washington, DC: National Academy of Medicine. PREPUBLICATION COPY - Uncorrected Proofs “Knowing is not enough; we must apply. Willing is not enough; we must do.” --GOETHE PREPUBLICATION COPY - Uncorrected Proofs ABOUT THE NATIONAL ACADEMY OF MEDICINE The National Academy of Medicine is one of three Academies constituting the Nation- al Academies of Sciences, Engineering, and Medicine (the National Academies). The Na- tional Academies provide independent, objective analysis and advice to the nation and conduct other activities to solve complex problems and inform public policy decisions.
    [Show full text]
  • Artificial Intelligence
    TechnoVision The Impact of AI 20 18 CONTENTS Foreword 3 Introduction 4 TechnoVision 2018 and Artificial Intelligence 5 Overview of TechnoVision 2018 14 Design for Digital 17 Invisible Infostructure 26 Applications Unleashed 33 Thriving on Data 40 Process on the Fly 47 You Experience 54 We Collaborate 61 Applying TechnoVision 68 Conclusion 77 2 TechnoVision 2018 The Impact of AI FOREWORD We introduce TechnoVision 2018, now in its eleventh year, with pride in reaching a second decade of providing a proven and relevant source of technology guidance and thought leadership to help enterprises navigate the compelling and yet complex opportunities for business. The longevity of this publication has been achieved through the insight of our colleagues, clients, and partners in applying TechnoVision as a technology guide and coupling that insight with the expert input of our authors and contributors who are highly engaged in monitoring and synthesizing the shift and emergence of technologies and the impact they have on enterprises globally. In this edition, we continue to build on the We believe that with TechnoVision 2018, you can framework that TechnoVision has established further crystalize your plans and bet on the right for several years with updates on last years’ technology disruptions, to continue to map and content, reflecting new insights, use cases, and traverse a successful digital journey. A journey technology solutions. which is not about a single destination, but rather a single mission to thrive in the digital epoch The featured main theme for TechnoVision 2018 through agile cycles of transformation delivering is AI, due to breakthroughs burgeoning the business outcomes.
    [Show full text]
  • The Deep Learning Revolution and Its Implications for Computer Architecture and Chip Design
    The Deep Learning Revolution and Its Implications for Computer Architecture and Chip Design Jeffrey Dean Google Research [email protected] Abstract The past decade has seen a remarkable series of advances in machine learning, and in particular deep learning approaches based on artificial neural networks, to improve our abilities to build more accurate systems across a broad range of areas, including computer vision, speech recognition, language translation, and natural language understanding tasks. This paper is a companion paper to a keynote talk at the 2020 International Solid-State Circuits Conference (ISSCC) discussing some of the advances in machine learning, and their implications on the kinds of computational devices we need to build, especially in the post-Moore’s Law-era. It also discusses some of the ways that machine learning may also be able to help with some aspects of the circuit design process. Finally, it provides a sketch of at least one interesting direction towards much larger-scale multi-task models that are sparsely activated and employ much more dynamic, example- and task-based routing than the machine learning models of today. Introduction The past decade has seen a remarkable series of advances in machine learning (ML), and in particular deep learning approaches based on artificial neural networks, to improve our abilities to build more accurate systems across a broad range of areas [LeCun et al. 2015]. Major areas of significant advances ​ ​ include computer vision [Krizhevsky et al. 2012, Szegedy et al. 2015, He et al. 2016, Real et al. 2017, Tan ​ ​ ​ ​ ​ ​ and Le 2019], speech recognition [Hinton et al.
    [Show full text]
  • Ai & the Sustainable Development Goals
    1 AI & THE SUSTAINABLE DEVELOPMENT GOALS: THE STATE OF PLAY Global Goals Technology Forum 2 3 INTRODUCTION In 2015, 193 countries agreed to the United Nations (UN) 2030 Agenda for Sustainable Development, which provides a shared blueprint for peace and prosperity for people and the planet, now and into the future. At its heart are the 17 Sustainable Development Goals (SDGs), which are an urgent call for action by all countries – developed and developing – in a global partnership. Achieving the SDGs is not just a moral imperative, but an economic one. While the world is making progress in some areas, we are falling behind in delivering the SDGs overall. We need all actors – businesses, governments, academia, multilateral institutions, NGOs, and others – to accelerate and scale their efforts to deliver the SDGs, using every tool at their disposal, including artificial intelligence (AI). In December 2017, 2030Vision 2030Vision published its first report, Uniting to “WE ARE AT A PIVOTAL Deliver Technology for the Global ABOUT 2030VISION Goals, which addressed the role MOMENT. AS ARTIFICIAL of digital technology – big data, AI & The Sustainable Development Goals: The State of Play State Goals: The Development Sustainable AI & The Founded in 2017, 2030Vision INTELLIGENCE BECOMES robotics, internet of things, AI, and is a partnership of businesses, MORE WIDELY ADOPTED, other technologies – in achieving NGOs, and academia that aims to the SDGs. transform the use of technology to WE HAVE A TREMENDOUS support the delivery of the SDGs. OPPORTUNITY TO In this paper, we focus on AI for 2030Vision serves as a platform REEVALUATE WHAT WE’VE the SDGs.
    [Show full text]
  • AI Principles 2020 Progress Update
    AI Principles 2020 Progress update AI Principles 2020 Progress update Table of contents Overview ........................................................................................................................................... 2 Culture, education, and participation ....................................................................4 Technical progress ................................................................................................................... 5 Internal processes ..................................................................................................................... 8 Community outreach and exchange .....................................................................12 Conclusion .....................................................................................................................................17 Appendix: Research publications and tools ....................................................18 Endnotes .........................................................................................................................................20 1 AI Principles 2020 Progress update Overview Google’s AI Principles were published in June 2018 as a charter to guide how we develop AI responsibly and the types of applications we will pursue. This report highlights recent progress in AI Principles implementation across Google, including technical tools, educational programs and governance processes. Of particular note in 2020, the AI Principles have supported our ongoing work to address
    [Show full text]
  • User Experience + Research Scientist Collaboration in a Generative Machine Learning Interface
    CHI 2019 Case Study CHI 2019, May 4–9, 2019, Glasgow, Scotland, UK Identifying the Intersections: User Figure 1: Google AI. http://ai.google Experience + Research Scientist Collaboration in a Generative Machine Learning Interface Claire Kayacik, Sherol Chen, Signe Noerly, Jess Holbrook, Adam Roberts, Douglas Eck Google, Inc. Mountain View, CA , 94043 USA (cbalgemann,sherol,noerly,jessh,adarob,deck)@google.com ABSTRACT Creative generative machine learning interfaces are stronger when multiple actors bearing different points of view actively contribute to them. User experience (UX) research and design involvement in the creation of machine learning (ML) models help ML research scientists to more effectively identify human needs that ML models will fulfill. The People and AI Research (PAIR) group within Google developed a novel program method in which UXers are embedded into an ML research group for three Figure 2: Magenta is a project within Google Brain that focuses on training ma- months to provide a human-centered perspective on the creation of ML models. The first full-time chines for creative expressivity, genera- cohort of UXers were embedded in a team of ML research scientists focused on deep generative models tive personalization, and intuitive inter- to assist in music composition. Here, we discuss the structure and goals of the program, challenges we faces. https://g.co/magenta faced during execution, and insights gained as a result of the process. We offer practical suggestions Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page.
    [Show full text]
  • Google Cloud Platform: Healthcare Solutions Playbook
    Google Cloud Platform: Healthcare Solutions Playbook Cloud training, tools, and reference architectures for researchers - Prepared for the National Institutes of Health (NIH) STRIDES. Author(s):​ Adrienne Wilkins, Trenton Leslie Title:​ Technical Account Manager (TAM), Intern PROFESSIONAL SERVICES Contents Introduction 2 Learning 2 Data Management Solutions 4 Reference Architectures 5 Genomics and Secondary Analysis 5 Patient Monitoring 9 Variant Analysis 11 Healthcare API for Machine Learning and Analytics 14 Radiological Image Extraction 18 Appendix 21 Page 1 PROFESSIONAL SERVICES Introduction The National Institutes of Health (NIH) established The Science and Technology Research Infrastructure for Discovery, Experimentation, and Sustainability (STRIDES) initiative to provide biomedical researchers with access to advanced, cost-effective, cloud-based computational infrastructure, tools, and services. Through STRIDES, researchers can take advantage of emerging data management methodologies, technological expertise, computational platforms, and tools to support cutting-edge experimentation and innovation. NIH has partnered with Google Cloud to support the STRIDES initiative through cloud services. In support of STRIDES, we’ve developed sets of playbooks to help enable researchers to build healthcare and life sciences solutions on Google Cloud Platform (GCP). The goal of this playbook it to aid researchers in developing healthcare systems and workflows on GCP. This playbook will provide GCP architecture and solutions examples for implementing
    [Show full text]
  • Microsoft Outlook
    From: Scott Wilson <[email protected]> Sent: Wednesday, November 13, 2019 5:18 PM To: aipartnership Subject: Re: Request for Comments on IP protection for AI innovation Docket No. PTO- C-2019-0038 It appears that companies such as Google and others are using copyrighted material for machine learning training datasets and creating for-profit products and patents that are partially dependent on copyrighted works. For instance, Google AI is using YouTube videos for machine learning products, and they have posted this information on their GoogleAI blog on multiple occasions. There are Google Books ML training datasets available in multiple locations on the internet, including at Amazon.com. There are two types of machine learning processes that I am aware of - expressive and non-expressive. A non- expressive ML training dataset would catalogue how many times a keyword appears in a copyrighted text, and appears to be permitted by fair use. However, expressive ML training analyzing artistic characteristics of copyrighted works to create patented products and/or processes does not appear to be exempted by fair use copyright laws. The whole issue revolves around the differences between research and product/profit creation in a corporate context. I am including an article below that examines these legal issues in more detail. https://lawandarts.org/wp-content/uploads/sites/14/2017/12/41.1_Sobel-FINAL.pdf or www.bensobel.org - first article on page links to the link above. As a copyright holder, I am concerned that this is a widespread industry practice of using copyrighted material without the knowledge or consent of copyright holders to create for-profit AI products that not only exploits copyright creators, but increasingly separates them from the ability to profit from their own work.
    [Show full text]
  • Google Goes All in on Artificial Intelligence As 'Google Research' Becomes 'Google AI'
    AiA Art News-service Google Goes All In On Artificial Intelligence As 'Google Research' Becomes 'Google AI' Sam Shead , CONTRIBUTORI cover artificial intelligence and Google DeepMind. Opinions expressed by Forbes Contributors are their own. MOUNTAIN VIEW, CA - MAY 08: Google CEO Sundar Pichai delivers the keynote address at the Google I/O 2018 Conference at Shoreline Amphitheater on May 8, 2018 in Mountain View, California. Google's two day developer conference runs through Wednesday May 9. (Photo by Justin Sullivan/Getty Images) Google has rebranded the whole of its "Google Research" division to Google AI as the company aggressively pursues developments in the burgeoning field of artificial intelligence. The change, announced ahead of its Google I/O developers conference this week, shows just how serious Google is when it comes to AI, which encompasses technologies such as computer vision, deep learning, and speech recognition. Announcing the news in a blog post, Google said it has been implementing machine learning into nearly everything it does over the last few years. "To better reflect this commitment, we're unifying our efforts under "Google AI", which encompasses all the state-of-the-art research happening across Google," wrote Christian Howard, editor-in-chief of Google AI Communications. Google has changed the homepage for Google AI so that it features published research on a variety of subject matters, stories of AI in action, and open source material and tools that others can use. The old Google Research site redirects to the new Google AI site. Google has also renamed the Google Research Twitter and Google+ channels to Google AI as part of the change.
    [Show full text]
  • Presentation by Amr Gaber
    https://m.facebook.com/story.php?story_fbid=2191943484425083&id=231112080362037 Transcript of UN remarks by Amr Gaber at CCW 2018/28/8 [Starts around 17:40] Hello my name is Amr Gaber. I am a very low level Google software engineer and I'm here speaking on behalf of myself and not my employer. I am also a volunteer with the Tech Workers Coalition. We’re a labor organization of tech workers, community members and other groups that are focused on bringing tech workers together from across industries, companies and backgrounds in order to make technology and the tech industry more equitable and just. I'm here to talk to you today because I took action with several of my co-workers at Google to stop Google from pursuing artificial intelligence in military applications. So there is a contract called Project Maven which is also a larger Department of Defense project and Google's involvement was to use artificial intelligence to enhance drone surveillance and targeting through image recognition. So I heard about this project from my co-workers internally through whistleblowers who were sharing it in the company forums. I co-wrote an open letter with several other Google employees to the CEO Sundar Pichai asking Google to cancel the project. Over 4,000 of my co-workers signed it, but that alone was not enough. It was actually a combination of workers, academics, students and outside groups that eventually got Google to step back from the contract. Organizations like the Tech Workers Coalition, Coworker.org, the Campaign To Stop Killer Robots, and the International Committee for Robot Arms Control which wrote a petition that over 1,000 academics signed including Noam Chomsky supporting me and my co-workers and our call to cancel the contract.
    [Show full text]
  • AI Explainability Whitepaper
    AI Explainability Whitepaper About This Whitepaper This whitepaper is a technical reference accompanying Google Cloud's AI Explanations product. It is targeted at model developers & data scientists responsible for designing and delivering ML models. Our goal is for them to leverage AI Explanations to simplify model development as well as explain the model’s behavior to key stakeholders. Product managers, business leaders and end users may also find parts of this whitepaper relevant, particularly around use cases enabled by AI Explanations and, crucially, considerations around its proper usage as well as its current limitations. Specifically, we refer these readers to the "​Usage Examples​" as well as the "​Attribution Limitations and Usage Considerations​" sections. Overview Evolution of Machine Learning In the quest for more accurate AI, the availability of compute resources coupled with increasing dataset sizes have fueled a trend towards more complex non-linear models. Indeed, we've seen a progression from handcrafted rules and heuristics, to linear models and decision trees, ensembles and deep models to, most recently, meta-learning or models that create other models. 1 These more accurate models have resulted in a paradigm shift along multiple dimensions: - Expressiveness ​enables fitting a wide range of functions in an increasing number of domains like forecasting, ranking, autonomous driving, particle physics, drug discovery, etc. - Versatility​ unlocks data modalities (image, audio, speech, text, tabular, time series, etc.) and enables joint/multi-modal applications. - Adaptability ​to small data regimes through transfer and multi-task learning. - Efficiency​ through custom optimized hardware like GPUs and TPUs enabling practitioners to train bigger models faster and cheaper.
    [Show full text]
  • Data-Driven Mergers and Personalization∗
    Data-Driven Mergers and Personalization∗ Zhijun Chen,y Chongwoo Choe,z Jiajia Cong,x and Noriaki Matsushima{ October 27, 2020 First draft. Comments welcome. Abstract Recent years have seen growing cases of data-driven tech mergers such as Google/Fitbit, in which a dominant digital platform acquires a relatively small firm possessing a large volume of consumer data. The digital platform can consolidate the consumer data with its existing data set from other services and use it for personalization in related markets. We develop a theoretical model to examine the impact of such mergers across the two markets that are related through a consumption synergy. The merger links the markets for data collection and data application, through which the digital platform can leverage its market power and hurt competitors in both markets. Personalization can lead to exploitation of some consumers in the market for data application. But insofar as competitors remain active, the merger increases total consumer surplus in both markets by intensifying competition. When the consumption synergy is large enough, the merger can result in monopolization of both markets, leading to further consumer harm when stand-alone competitors exit in the long run. Thus there is a trade-off where potential dynamic costs can outweigh static benefits. We also discuss policy implications by considering various merger remedies. Key words: big data, personalization, tech mergers JEL Classification: D43, L13, L41, K21 ∗We are grateful to Arthur Campbell, Greg Crawford, Jacques Cr´emer,Doh-Shin Jeon, Bruno Jullien, Paul Heidhues, Vai-Lam Mui, Patrick Rey, Chengsi Wang, and Xiaojian Zhao for many helpful comments.
    [Show full text]