Handbook of Natural Language Processing and Machine Translation

Total Page:16

File Type:pdf, Size:1020Kb

Handbook of Natural Language Processing and Machine Translation Handbook of Natural Language Processing and Machine Translation Joseph Olive · Caitlin Christianson · John McCary Editors Handbook of Natural Language Processing and Machine Translation DARPA Global Autonomous Language Exploitation 123 Editors Joseph Olive Caitlin Christianson Defense Advanced Research Projects Agency Defense Advanced Research Projects Agency IPTO Reston Virginia, USA N Fairfax Drive 3701 [email protected] Arlington, VA 22203, USA [email protected] John McCary Defense Advanced Research Projects Agency Bethesda Maryland, USA [email protected] ISBN 978-1-4419-7712-0 e-ISBN 978-1-4419-7713-7 DOI 10.1007/978-1-4419-7713-7 Springer New York Dordrecht Heidelberg London Library of Congress Control Number: 2011920954 c Springer Science+Business Media, LLC 2011 All rights reserved. This work may not be translated or copied in whole or in part without the written permission of the publisher (Springer Science+Business Media, LLC, 233 Spring Street, New York, NY 10013, USA), except for brief excerpts in connection with reviews or scholarly analysis. Use in connection with any form of information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now known or hereafter developed is forbidden. The use in this publication of trade names, trademarks, service marks, and similar terms, even if they are not identified as such, is not to be taken as an expression of opinion as to whether or not they are subject to proprietary rights. Printed on acid-free paper Springer is part of Springer Science+Business Media (www.springer.com) Acknowledgements First, I would like to thank the community for all of its hard work in making GALE a success. I would like to thank the technical assistants Caitlin Christianson and John McCary for all of their help with the program and this book. Special thanks to Paul Dietrich for making GALE run smoothly. We would like to thank the current DARPA management for its continued support throughout the program, especially DARPA Director Regina Dugan, Deputy DARPA Director Ken Gabriel, Information Processing Techniques Office (IPTO) Director Daniel Kaufman, and IPTO Deputy Director Mark Luettgen. We would also like to thank previous IPTO Directors and Deputy Directors Ron Brachman, Charles Holland, Barbara Yoon, Charles Holland, and Charles Morefield for their help in launching the program, continued support, and encouragement to write this book. Special thanks to former DARPA Director Anthony Tether for having the vision and the faith in us to fund the program so that it had a chance to succeed. Finally, I would like to thank my wife Virginia Kayhart Olive. Without her great personal sacrifice this project would not have been possible. v Introduction Authors: Joseph Olive, Caitlin Christianson, and John McCary “When I use a word, it means just what I choose it to mean – neither more nor less.” Humpty Dumpty in Lewis Carroll’s Through the Looking Glass, or Alice in Wonderland The meaning of what Lewis Carroll’s Humpty Dumpty says is abundantly clear to him, but to others, it is virtually incomprehensible. Because the rules of his language are entirely of his own devising, he is the only one who knows what he means. Translation of one language into another poses a similar problem; knowing the usual meaning of a word is not enough. To translate, it is necessary to convey the meaning of the entire message, not just transfer words from one language to another. Because people can perform this task so adeptly, it is easy to underestimate the challenge it poses to computers. Although computational capabilities of machines exceed those of humans in many ways, even the most advanced of today’s computers cannot match the language ability that humans acquire naturally. To translate and extract information conveyed through language, humans take advantage of a variety of cognitive abilities that no computer can currently emulate. The Defense Advanced Research Projects Agency (DARPA), however, specializes in tackling just such challenging problems. DARPA researchers have attacked the problem of machine translation as part of the Global Autonomous Language Exploitation (GALE) Program. Like other programs at DARPA, GALE was initiated to fill a need for the Defense Department. In the case of GALE, this need is to close the language gap – to make relevant information, regardless of source language, accessible to English-speaking personnel. The program’s goal is to create technology to automatically translate, analyze, and distill the information that forms the overwhelming tidal wave of foreign language material facing Defense Department personnel. The value of GALE technology is to be creation of the ability not only to translate, but also to identify relevant information and sort it from what is not, and to present a final product in a form both understandable and useable. While there is no existing parallel for such a capability, there have certainly been fictionalized precedents to the idea behind GALE – the universal translator, capable of translating between English and thousands of other languages, in the form of a compact silver device worn on the chest of every Star Trek crew member; the HAL 9000 computer, capable of reasoning and defending itself in 2001, Stanley Kubrick’s classic vii viii Introduction futuristic film; and Star Wars’ robot C3PO, who speaks nearly every known language in the universe. For the purpose of research and development of automated translation and human language processing capability in the GALE Program, language has been classified into two input modes– speech and electronic text. A third important language input mode, hardcopy document text, is the subject of another DARPA program. While research related to each of GALE’s two input modes focuses on producing a correct translation and extracting information, each mode of input presents singular problems that require different research paths. Unlike text, which generally has orthographic separation, speech signals are continuous, lacking word and phrase boundary markers. In speech, even sentence boundaries are difficult to determine. In addition, the confusability of many phonemes adds to the uncertainty. Because these difficulties are alleviated when a computer carries out transcription using a more explicit orthographic system than those used by human writers, machine translation researchers previously attempted to use computers to transcribe speech into text and then translate that text. In the GALE Program, researchers have begun to combine the processes of transcription and translation, enabling information about possible translations to assist transcription, and information about possible transcriptions and transcription ambiguities to assist translation. An important benefit of using this method is that interactivity between transcription and translation reduces instances of propagation of errors made earlier in the process by providing opportunities for correction. This interactive technique has yielded significant improvements in accuracy for both transcription and translation of speech. In this way, GALE researchers have achieved revolutionary progress by consistently and effectively blending previously distinct speech and text technologies. Text input can also be problematic. In many languages, word boundaries are clear in writing, but due to the lack of orthographic representation of other language elements, such as prosody, it can still be difficult to know a writer’s intent. Scripts without word boundaries introduce additional uncertainty in reading. Chinese writing, for example, does not indicate word boundaries orthographically, a characteristic that can create ambiguity. Semitic language scripts do indicate word boundaries, but often do not include explicit vowel marking, thus creating ambiguities, since it can be uncertain which vowels were intended. GALE researchers have undertaken extraordinary efforts to address these and other obstacles in machine translation of text. One of the greatest challenges in planning an approach for GALE was defining precisely what tasks GALE’s natural language processing machines would be achieving. Was it the ability to translate any language into English? Or was there an even higher goal of retrieving what was relevant from the input? Would achieving such a goal mean that GALE researchers would have to create technology that could extract relevant information from translated material and operate on foreign language material directly? Would GALE machines be able to perform all of these tasks well enough to enable assessment and analysis of the volume of information now available to anyone connected to the Internet or satellite television? These questions have resulted in many challenges to GALE and refinements of the program’s fundamental aspects. Handbook of Natural Language Processing and Machine Translation ix A Partial History of Human Language Technology Research at DARPA Author: Allen Sears During the past four decades, DARPA has sponsored a wide variety of research on human language technology – efforts that turned out to be stepping stones to GALE. DARPA entered the speech recognition field in 1971 with the launch of the five-year Speech Understanding Research (SUR) Program. Although its immediate impact was limited, SUR included pioneering work with hidden Markov models, which lie at the heart of all modern speech-to-text systems. DARPA speech and text processing research proceeded at a relatively
Recommended publications
  • Part 5: Machine Translation Evaluation Editor: Bonnie Dorr
    Part 5: Machine Translation Evaluation Editor: Bonnie Dorr Chapter 5.1 Introduction Authors: Bonnie Dorr, Matt Snover, Nitin Madnani The evaluation of machine translation (MT) systems is a vital field of research, both for determining the effectiveness of existing MT systems and for optimizing the performance of MT systems. This part describes a range of different evaluation approaches used in the GALE community and introduces evaluation protocols and methodologies used in the program. We discuss the development and use of automatic, human, task-based and semi-automatic (human-in-the-loop) methods of evaluating machine translation, focusing on the use of a human-mediated translation error rate HTER as the evaluation standard used in GALE. We discuss the workflow associated with the use of this measure, including post editing, quality control, and scoring. We document the evaluation tasks, data, protocols, and results of recent GALE MT Evaluations. In addition, we present a range of different approaches for optimizing MT systems on the basis of different measures. We outline the requirements and specific problems when using different optimization approaches and describe how the characteristics of different MT metrics affect the optimization. Finally, we describe novel recent and ongoing work on the development of fully automatic MT evaluation metrics that can have the potential to substantially improve the effectiveness of evaluation and optimization of MT systems. Progress in the field of machine translation relies on assessing the quality of a new system through systematic evaluation, such that the new system can be shown to perform better than pre-existing systems. The difficulty arises in the definition of a better system.
    [Show full text]
  • Joshua 2.0: a Toolkit for Parsing-Based Machine Translation with Syntax, Semirings, Discriminative Training and Other Goodies
    Joshua 2.0: A Toolkit for Parsing-Based Machine Translation with Syntax, Semirings, Discriminative Training and Other Goodies Zhifei Li, Chris Callison-Burch, Chris Dyer,y Juri Ganitkevitch, Ann Irvine, Lane Schwartz,? Wren N. G. Thornton, Ziyuan Wang, Jonathan Weese and Omar F. Zaidan Center for Language and Speech Processing, Johns Hopkins University, Baltimore, MD y Computational Linguistics and Information Processing Lab, University of Maryland, College Park, MD ? Natural Language Processing Lab, University of Minnesota, Minneapolis, MN Abstract authors and several other groups in their daily re- search, and has been substantially refined since the We describe the progress we have made in first release. The most important new functions in the past year on Joshua (Li et al., 2009a), the toolkit are: an open source toolkit for parsing based • Support for any style of synchronous context machine translation. The new functional- free grammar (SCFG) including syntax aug- ity includes: support for translation gram- ment machine translation (SAMT) grammars mars with a rich set of syntactic nonter- (Zollmann and Venugopal, 2006) minals, the ability for external modules to posit constraints on how spans in the in- • Support for external modules to posit transla- put sentence should be translated, lattice tions for spans in the input sentence that con- parsing for dealing with input uncertainty, strain decoding (Irvine et al., 2010) a semiring framework that provides a uni- fied way of doing various dynamic pro- • Lattice parsing for dealing with
    [Show full text]
  • TER-Plus: Paraphrase, Semantic, and Alignment Enhancements to Translation Edit Rate
    Mach Translat DOI 10.1007/s10590-009-9062-9 TER-Plus: paraphrase, semantic, and alignment enhancements to Translation Edit Rate Matthew G. Snover · Nitin Madnani · Bonnie Dorr · Richard Schwartz Received: 15 May 2009 / Accepted: 16 November 2009 © Springer Science+Business Media B.V. 2009 Abstract This paper describes a new evaluation metric, TER-Plus (TERp) for auto- matic evaluation of machine translation (MT). TERp is an extension of Translation Edit Rate (TER). It builds on the success of TER as an evaluation metric and alignment tool and addresses several of its weaknesses through the use of paraphrases, stemming, synonyms, as well as edit costs that can be automatically optimized to correlate better with various types of human judgments. We present a correlation study comparing TERptoBLEU,METEOR and TER, and illustrate that TERp can better evaluate translation adequacy. Keywords Machine translation evaluation · Paraphrasing · Alignment M. G. Snover (B) · N. Madnani · B. Dorr Laboratory for Computational Linguistics and Information Processing, Institute for Advanced Computer Studies, University of Maryland, College Park, MD, USA e-mail: [email protected] N. Madnani e-mail: [email protected] B. Dorr e-mail: [email protected] R. Schwartz BBN Technologies, Cambridge, MA, USA e-mail: [email protected] 123 M. G. Snover et al. 1 Introduction TER-Plus, or TERp1 (Snover et al. 2009), is an automatic evaluation metric for machine translation (MT) that scores a translation (the hypothesis) of a foreign lan- guage text (the source) against a translation of the source text that was created by a human translator, which we refer to as a reference translation.
    [Show full text]
  • The History and Promise of Machine Translation
    The History and Promise of Machine Translation Lane Schwartz Department of Linguistics University of Illinois at Urbana-Champaign Urbana IL, USA [email protected] Abstract This work examines the history of machine translation (MT), from its intellectual roots in the 17th century search for universal language through its practical realization in the late 20th and early 21st centuries. We survey the major MT paradigms, including transfer-based, interlingua, and statistical approaches. We examine the current state of human-machine partnership in translation, and consider the substantial, yet largely unfulfilled, promise that MT technology has for human translation professionals. 1 1 Introduction With the possible exception of the calculation of artillery trajectory tables (Goldstine and Goldstine 1946), machine translation has a strong claim to be the oldest established research discipline within computer science. Machine translation is a modern discipline, one whose success and very existence is predicated on the existence of modern computing hardware. Yet certain core ideas which would eventually serve as the foundations for this new discipline have roots which predate the development of electronic digital computers. This work examines the history of machine translation (MT), from its intellectual roots in the 17th century search for universal language through its practical realization in the late 20th and early 21st centuries. Beginning with the development of the first general-purpose electronic digital computers in the late 1940s, substantial
    [Show full text]
  • CURRICULUM Vitae Nitin MADNANI
    CURRICULUM vitae Nitin MADNANI 666 Rosedale Rd., MS 13-R [email protected] Princeton NJ 08541 http://www.desilinguist.org Current Status: Legal Permanent Resident Current Senior Research Scientist, NLP & Speech Position Educational Testing Service (ETS), Princeton, New Jersey (2017-present) Previous Research Scientist, NLP & Speech Positions Educational Testing Service (ETS), Princeton, New Jersey (2013-2017) Associate Research Scientist, NLP & Speech Educational Testing Service (ETS), Princeton, New Jersey (2010-2013) Education University of Maryland, College Park, MD. Ph.D. in Computer Science, GPA: 4.0 Dissertation Title: The Circle of Meaning: From Translation to Paraphrasing and Back 2010. University of Maryland, College Park, MD. M.S in Computer Engineering, GPA: 3.77 Concentration: Computer Organization, Microarchitecture and Embedded Systems 2004. Punjab Engineering College, Panjab University, India. B.E. in Electrical Engineering, With Honors Senior Thesis: Interactive Visualization of Grounding Systems for Power Stations 2000. Professional Computational Linguistics, Natural Language Processing, Statistical Machine Translation, Auto- Interests matic Paraphrase Generation, Machine Learning, Artificial Intelligence, Educational Technology, and Computer Science Education. Publications Book Chapters ◦ Jill Burstein, Beata Beigman-Klebanov, Nitin Madnani and Adam Faulkner. Sentiment Analy- sis & Detection for Essay Evaluation. Handbook for Automated Essay Scoring, Taylor and Francis. Mark D. Shermis & Jill Burstein (eds.). 2013. ◦ Jill Burstein, Joel Tetreault and Nitin Madnani. The E-rater Automated Essay Scoring System. Handbook for Automated Essay Scoring, Taylor and Francis. Mark D. Shermis & Jill Burstein (eds.). 2013. ◦ Yaser Al-Onaisan, Bonnie Dorr, Doug Jones, Jeremy Kahn, Seth Kulick, Alon Lavie, Gregor Leusch, Nitin Madnani, Chris Manning, Arne Mauser, Alok Parlikar, Mark Przybocki, Rich Schwartz, Matt Snover, Stephan Vogel and Clare Voss.
    [Show full text]