Efficient Language Modeling Algorithms with Applications To

Total Page:16

File Type:pdf, Size:1020Kb

Efficient Language Modeling Algorithms with Applications To Efficient Language Modeling Algorithms with Applications to Statistical Machine Translation Kenneth Heafield CMU-LTI-13-017 Language Technologies Institute School of Computer Science Carnegie Mellon University 5000 Forbes Ave., Pittsburgh, PA 15213 www.lti.cs.cmu.edu Thesis Committee: Alon Lavie, Chair Chris Dyer Bhiksha Raj Philipp Koehn, University of Edinburgh Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy In Language and Information Technologies Copyright c , 2013 Kenneth Heafield Abstract N-gram language models are an essential component in statistical natural language processing systems for tasks such as machine translation, speech recognition, and optical character recognition. They are also re- sponsible for much of the computational costs. This thesis contributes efficient algorithms for three language modeling problems: estimating probabilities from corpora, representing a model in memory, and searching for high-scoring output when log language model probability is part of the score. Most existing language modeling toolkits operate in RAM, effectively limiting model size. This work contributes disk-based streaming algorithms that use a configurable amount of RAM to estimate Kneser- Ney language models 7.13 times as fast as the popular SRILM toolkit. Scaling to 126 billion tokens led to first-place performance in the 2013 Workshop on Machine Translation for all three language pairs where submissions were made. Query speed is critical because a machine translation system makes millions of queries to translate one sentence. Thus, language models are typically queried in RAM, where size is a concern. This work contributes two near-lossless data structures for efficient storage and querying. The first, based on linear probing hash tables, responds to queries 2.42 times as fast as the SRILM toolkit while using 57% of the memory. The second, based on sorted arrays, is faster than all baselines and uses less memory than all lossless baselines. Searching for high-scoring output is difficult because log language model probabilities do not sum when strings are concatenated. This thesis contributes a series of optimizations that culminate in a new approx- imate search algorithm. The algorithm applies to search spaces expressed as lattices and, more generally, hypergraphs that arise in many natural language tasks. Experiments with syntactic machine translation show that the new algorithm attains various levels of accuracy 3.25 to 10.01 times as fast as the popular cube pruning algorithm with SRILM. Acknowledgments First, I would like to thank my adviser, Alon Lavie, who is primarily responsible for introducing me to machine translation. He especially deserves thanks for supporting a side project on querying language models; that project eventually grew in scope to become this thesis. Philipp Koehn hosted me at the University of Edinburgh for more than a year and never turned down a travel request. It is because of him that I was able to focus for the final year of this work. He is also one of the organizers of MT Marathon, where some of the chapters started. It is rare to have a committee that follows—and uses—the work as it is being performed. Chris Dyer provided detailed feedback, encouragement, and help with cdec. He has also been an early adopter and unwitting tester. Bhiksha Raj provided insight into language modeling in general and for speech recognition, including and beyond his own work on the subject. My colleagues at Carnegie Mellon contributed knowledge and discussions. Jonathan Clark was espe- cially helpful with estimating language models and came up with the name lmplz. Greg Hanneman deserves special mention for contributing his French–English system for use in experiments and answering questions about it. Michael Denkowski has been helpful both with METEOR and evaluations. Victor Chahuneau contributed ideas and code. Thanks to Nathan Schneider, Narges Razavian, Kevin Gimpel, Yubin Kim and the tea group for making it entertaining, even though I rarely drink tea. Kami Vaniea has been especially supportive. The University of Edinburgh became my adoptive research family. Hieu Hoang motivated me to improve my work and release it to a broader audience. In addition to helping with Moses, he is to credit for the name KenLM. Barry Haddow has been a great resource on the inner workings of Moses. Hieu Hoang, Barry Haddow, Hervé Saint-Amand, and Ulrich Germann were all supportive in research and as officemates. Nadir Durrani’s collaboration was essential to the submissions to the Workshop on Machine Translation. Miles Osborne provided his preprocessed copy of the ClueWeb09 corpus and encouraged me to think big with language modeling. Alexandra Birch, Christian Buck, Liane Guillou, Eva Hasler, and Phil Williams have all been helpful in discussions. MT Marathon is a great venue to work with other researchers. Hieu Hoang, Tetsuo Kiso and Marcello Federico joined my project on state minimization that eventually grew into a chapter. The next year, Ivan Pouzyrevsky and Mohammed Mediani joined my project on estimating language models. Ivan has been particularly helpful with optimizing disk-based algorithms. System administators from Carnegie Mellon (especially Michael Stroucken), the Texas Advanced Com- puting Center (TACC) at The University of Texas at Austin, the San Diego Supercomputer Center, and the Unversity of Edinburgh (especially Hervé Saint-Amand and Nicholas Moir) all deserve thanks for handling my requests, such as increasing the maximum size of RAM disks to 700 GB on the machines with 1 TB RAM. Finally, my users are a source of motivation, ideas, and code. Matt Post, Juri Ganitkevitch, Stephan Peitz, and Jörn Wübker, Jonathan Graehl, and Baskaran Sankaran have all been helpful in discussions and in increasing the impact of this work. i This work used the Extreme Science and Engineering Discovery Environment (XSEDE), which is sup- ported by National Science Foundation grant number OCI-1053575, specifically Stampede and Trestles un- der allocation TG-CCR110017. The author acknowledges the Texas Advanced Computing Center (TACC) at The University of Texas at Austin for providing HPC resources that have contributed to the research re- sults reported within this thesis. This work made use of the resources provided by the Edinburgh Compute and Data Facility (http://www.ecdf.ed.ac.uk/). The ECDF is partially supported by the eDIKT initiative (http://www.edikt.org.uk/). The research leading to these results has received fund- ing from the European Union Seventh Framework Programme (FP7/2007-2013) under grant agreements 287658 (EU BRIDGE), 287576 (CASMACAT), 287688 (MateCat), and 288769 (ACCEPT). This material is based upon work supported in part by the National Science Foundation Graduate Research Fellowship under Grant No. 0750271, by the National Science Foundation under grant IIS-0713402, by a NPRP grant (NPRP 09-1140-1-177) from the Qatar National Research Fund (a member of the Qatar Foundation), and by the DARPA GALE program. ii Contents 1 Introduction 1 1.1 Research Contributions.....................................4 1.2 Preview of Translation Results.................................5 2 Background on Language Models and Decoding7 2.1 N–gram Models.........................................8 2.2 Smoothing............................................8 2.3 Querying............................................ 10 2.4 Decoding............................................ 11 2.4.1 Decoding Problems................................... 11 2.4.2 Searching with Language Models........................... 15 2.4.3 Implementations.................................... 19 2.5 Summary............................................ 20 3 Estimating Kneser-Ney Language Models 23 3.1 Introduction........................................... 23 3.2 Related Work.......................................... 24 3.2.1 Google......................................... 24 3.2.2 Microsoft........................................ 25 3.2.3 SRI........................................... 25 3.2.4 IRST.......................................... 25 3.2.5 MIT........................................... 26 3.2.6 Berkeley........................................ 26 3.3 Estimation Pipeline....................................... 26 3.3.1 Counting........................................ 26 3.3.2 Adjusting Counts.................................... 27 3.3.3 Normalization..................................... 28 3.3.4 Interpolation...................................... 29 3.3.5 Joining......................................... 30 3.4 Numerical Precision....................................... 30 3.5 Streaming and Sorting Framework............................... 32 3.6 Experiments........................................... 34 3.6.1 Methodology...................................... 34 3.6.2 Memory Setting.................................... 34 3.6.3 Toolkit Comparison.................................. 34 3.6.4 Scaling......................................... 34 3.7 Summary............................................ 39 iii 4 Querying Language Models 41 4.1 Introduction........................................... 41 4.2 Data Structures......................................... 43 4.2.1 Hash Tables and Probing................................ 43 4.2.2 Sorted Arrays and Trie................................. 45 4.3 Related Work.......................................... 47 4.4 Threading and Memory Mapping................................ 48 4.5 Experiments..........................................
Recommended publications
  • (Meta-) Evaluation of Machine Translation
    (Meta-) Evaluation of Machine Translation Chris Callison-Burch Cameron Fordyce Philipp Koehn Johns Hopkins University CELCT University of Edinburgh ccb cs jhu edu fordyce celct it pkoehn inf ed ac uk Christof Monz Josh Schroeder Queen Mary, University of London University of Edinburgh christof dcs qmul ac uk j schroeder ed ac uk Abstract chine translation evaluation workshop. Here we ap- ply eleven different automatic evaluation metrics, This paper evaluates the translation quality and conduct three different types of manual evalu- of machine translation systems for 8 lan- ation. guage pairs: translating French, German, Beyond examining the quality of translations pro- Spanish, and Czech to English and back. duced by various systems, we were interested in ex- We carried out an extensive human evalua- amining the following questions about evaluation tion which allowed us not only to rank the methodologies: How consistent are people when different MT systems, but also to perform they judge translation quality? To what extent do higher-level analysis of the evaluation pro- they agree with other annotators? Can we im- cess. We measured timing and intra- and prove human evaluation? Which automatic evalu- inter-annotator agreement for three types of ation metrics correlate most strongly with human subjective evaluation. We measured the cor- judgments of translation quality? relation of automatic evaluation metrics with This paper is organized as follows: human judgments. This meta-evaluation re- veals surprising facts about the most com- • Section 2 gives an overview of the shared task. monly used methodologies. It describes the training and test data, reviews the baseline system, and lists the groups that 1 Introduction participated in the task.
    [Show full text]
  • Public Project Presentation and Updates
    7.2: Public Project Presentation and Updates Philipp Koehn Distribution: Public EuroMatrix Statistical and Hybrid Machine Translation Between All European Languages IST 034291 Deliverable 7.2 December 21, 2007 Project funded by the European Community under the Sixth Framework Programme for Research and Technological Development. Project ref no. IST-034291 Project acronym EuroMatrix Project full title Statistical and Hybrid Machine Translation Between All Eu- ropean Languages Instrument STREP Thematic Priority Information Society Technologies Start date / duration 01 September 2006 / 30 Months Distribution Public Contractual date of delivery September 1, 2007 Actual date of delivery September 13, 2007 Deliverable number 7.2 Deliverable title Public Project Presentation and Updates Type Report, contractual Status & version Finished Number of pages 19 Contributing WP(s) WP7 WP / Task responsible WP7 / 7.6 Other contributors none Author(s) Philipp Koehn EC project officer Xavier Gros Keywords The partners in EuroMatrix are: Saarland University (USAAR) University of Edinburgh (UEDIN) Charles University (CUNI-MFF) CELCT GROUP Technologies MorphoLogic For copies of reports, updates on project activities and other EuroMatrix-related in- formation, contact: The EuroMatrix Project Co-ordinator Prof. Hans Uszkoreit Universit¨at des Saarlandes, Computerlinguistik Postfach 15 11 50 66041 Saarbr¨ucken, Germany [email protected] Phone +49 (681) 302-4115- Fax +49 (681) 302-4700 Copies of reports and other material can also be accessed via the project’s homepage: http://www.euromatrix.net/ c 2007, The Individual Authors No part of this document may be reproduced or transmitted in any form, or by any means, electronic or mechanical, including photocopy, recording, or any information storage and retrieval system, without permission from the copyright owner.
    [Show full text]
  • 8.14: Annual Public Report
    8.14: Annual Public Report S. Busemann, O. Bojar, C. Callison-Burch, M. Cettolo, R. Garabik, J. van Genabith, P. Koehn, H. Schwenk, K. Simov, P. Wolf Distribution: Public EuroMatrixPlus Bringing Machine Translation for European Languages to the User ICT 231720 Deliverable 8.14 Project funded by the European Community under the Seventh Framework Programme for Research and Technological Development. Project ref no. ICT-231720 Project acronym EuroMatrixPlus Project full title Bringing Machine Translation for European Languages to the User Instrument STREP Thematic Priority ICT-2007.2.2 Cognitive systems, interaction, robotics Start date / duration 01 March 2009 / 38 Months Distribution Public Contractual date of delivery November 15, 2011 Actual date of delivery December 05, 2011 Date of last update December 5, 2011 Deliverable number 8.14 Deliverable title Annual Public Report Type Report Status & version Final Number of pages 13 Contributing WP(s) all WP / Task responsible DFKI Other contributors All Participants Author(s) S. Busemann, O. Bojar, C. Callison-Burch, M. Cettolo, R. Garabik, J. van Genabith, P. Koehn, H. Schwenk, K. Simov, P. Wolf EC project officer Michel Brochard Keywords The partners in DFKI GmbH, Saarbr¨ucken (DFKI) EuroMatrixPlus University of Edinburgh (UEDIN) are: Charles University (CUNI-MFF) Johns Hopkins University (JHU) Fondazione Bruno Kessler (FBK) Universit´edu Maine, Le Mans (LeMans) Dublin City University (DCU) Lucy Software and Services GmbH (Lucy) Central and Eastern European Translation, Prague (CEET) Ludovit Stur Institute of Linguistics, Slovak Academy of Sciences (LSIL) Institute of Information and Communication Technologies, Bulgarian Academy of Sciences (IICT-BAS) For copies of reports, updates on project activities and other EuroMatrixPlus-related information, contact: The EuroMatrixPlus Project Co-ordinator Prof.
    [Show full text]
  • Statistical and Hybrid Machine Translation Between All European Languages
    Statistical and hybrid machine translation between all European languages Publishable executive summary 01 September 2006 – 30 November 2007 Acronym: EuroMatrix Contract N°: IST 034291 Duration: 30 months from 01/09/2006 Coordinator: Prof. Dr. Hans Uszkoreit Saarland University D-66123 Saarbrücken [email protected] Scientific coordinator: Dr. Philipp Koehn Edinburgh University [email protected] PROJECT FUNDED BY THE EUROPEAN COMMUNITY UNDER THE SIXTH FRAMEWORK PROGRAMME FOR RESEARCH AND TECHNOLOGICAL DEVELOPMENT. Contractors involved Saarland University Germany Computational Linguistics The University of Edinburgh United Kingdom School of Informatics Univerzita Karlova v Praze/ Charles University Prague Czech Republic Institute of Formal and Applied Linguistics CELCT Center for the Evaluation of Language and Italy Communication Technologies MorphoLogic Hungary Germany Group Technologies AG -------------------------------------------------------------------------------------------- Publishable Executive Summary ©The EuroMatrix consortium, 2008 Summary description of project objectives EuroMatrix aims at a major push in machine translation technology applying the most advanced MT technologies systematically to all pairs of EU languages. Special attention is being paid to the languages of the new and near-term prospective member states. As part of this application development, EuroMatrix designs and investigates novel combinations of statistical techniques and linguistic knowledge sources as well as hybrid MT architectures. EuroMatrix addresses urgent European economic and social needs by concentrating on European languages and on high-quality translation to be employed for the publication of technical, social, legal and political documents. EuroMatrix aims at enriching the statistical MT approach with novel learning paradigms and experiment with new combinations of methods and resources from statistical MT, rule-based MT, shallow language processing and computational lexicography/morphology.
    [Show full text]
  • Abstract the Circle of Meaning: from Translation
    ABSTRACT Title of dissertation: THE CIRCLE OF MEANING: FROM TRANSLATION TO PARAPHRASING AND BACK Nitin Madnani, Doctor of Philosophy, 2010 Dissertation directed by: Professor Bonnie Dorr Department of Computer Science The preservation of meaning between inputs and outputs is perhaps the most ambitious and, often, the most elusive goal of systems that attempt to process natural language. Nowhere is this goal of more obvious importance than for the tasks of machine translation and paraphrase generation. Preserving meaning between the input and the output is paramount for both, the monolingual vs bilingual distinction notwithstanding. In this thesis, I present a novel, symbiotic relationship between these two tasks that I term the “circle of meaning”. Today’s statistical machine translation (SMT) systems require high quality human translations for parameter tuning, in addition to large bi-texts for learning the translation units. This parameter tuning usually involves generating translations at different points in the parameter space and obtaining feedback against human- authored reference translations as to how good the translations. This feedback then dictates what point in the parameter space should be explored next. To measure this feedback, it is generally considered wise to have multiple (usually 4) reference translations to avoid unfair penalization of translation hypotheses which could easily happen given the large number of ways in which a sentence can be translated from one language to another. However, this reliance on multiple reference translations creates a problem since they are labor intensive and expensive to obtain. Therefore, most current MT datasets only contain a single reference. This leads to the problem of reference sparsity—the primary open problem that I address in this dissertation— one that has a serious effect on the SMT parameter tuning process.
    [Show full text]
  • Improving Statistical Machine Translation Efficiency by Triangulation
    Improving Statistical Machine Translation Efficiency by Triangulation Yu Chen, Andreas Eisele, Martin Kay Saarland University Saarbrucken,¨ Germany {yuchen, eisele, kay}@coli.uni-sb.de Abstract In current phrase-based Statistical Machine Translation systems, more training data is generally better than less. However, a larger data set eventually introduces a larger model that enlarges the search space for the decoder, and consequently requires more time and more resources to translate. This paper describes an attempt to reduce the model size by filtering out the less probable entries based on testing correlation using additional training data in an intermediate third language. The central idea behind the approach is triangulation, the process of incorporating multilingual knowledge in a single system, which eventually utilizes parallel corpora available in more than two languages. We conducted experiments using Europarl corpus to evaluate our approach. The reduction of the model size can be up to 70% while the translation quality is being preserved. 1. Introduction 2. Phrase-based SMT The dominant paradigm in SMT is referred to as phrase- Statistical machine translation (SMT) is now generally based machine translation. The term refers to a sequence of taken to be an enterprise in which machine learning tech- words characterized by its statistical rather then any gram- niques are applied to a bilingual corpus to produce a trans- matical properties. At the center of the process is a phrase lation system entirely automatically. Such a scheme has table, a list of phrases identified in a source sentence to- many potential advantages over earlier systems which re- gether with potential translations.
    [Show full text]
  • Factored Translation Models and Discriminative Training For
    Machine Translation at Edinburgh Factored Translation Models and Discriminative Training Philipp Koehn, University of Edinburgh 9 July 2007 Philipp Koehn, University of Edinburgh EuroMatrix 9 July 2007 1 Overview • Intro: Machine Translation at Edinburgh • Factored Translation Models • Discriminative Training Philipp Koehn, University of Edinburgh EuroMatrix 9 July 2007 2 The European Challenge Many languages • 11 official languages in EU-15 • 20 official languages in EU-25 • many more minority languages Challenge • European reports, meetings, laws, etc. • develop technology to enable use of local languages as much as possible Philipp Koehn, University of Edinburgh EuroMatrix 9 July 2007 3 Existing MT systems for EU languages [from Hutchins, 2005] Cze Dan Dut Eng Est Fin Fre Ger Gre Hun Ita Lat Lit Mal Pol Por Slo Slo Spa Swe Czech –..1..11..1.........4 Danish .–.....1............1 Dutch ..–6..21............9 English 2 . 6 – . 42 48 3 3 29 1 . 7 30 2 . 48 1 222 Estonian ....–...............0 Finnish ...2.–.1............3 French 1 . 2 38 . – 22 3 . 9 . 1 5 . 10 . 91 German 1 1 1 49 . 1 23 – . 1 8 . 4 3 2 . 8 1 103 Greek ...2..3.–...........5 Hungarian ...1...1.–..........2 Italian 1 . 25 . 9 8 . – . 1 3 . 7 . 54 Latvian ...1.......–........1 Lithuanian ............–.......0 Maltese .............–......0 Polish ...6..13..1...–2..1.14 Portuguese . 25 . 4 4 . 3 . 1 – . 6 . 43 Slovak ...1...1........–...2 Slovene .................–..0 Spanish 1 . 42 . 8 7 . 7 . 1 6 . – . 72 Swedish ...2...1...........–3 6 1 9 201 0 1 93 99 6 4 58 1 0 0 15 49 4
    [Show full text]
  • Convergence of Translation Memory and Statistical Machine Translation
    Convergence of Translation Memory and Statistical Machine Translation Philipp Koehn Jean Senellart University of Edinburgh Systran 10 Crichton Street La Grande Arche Edinburgh, EH8 9AB 1, Parvis de la Defense´ Scotland, United Kingdom 92044 Paris, France [email protected] [email protected] Abstract tems (Koehn, 2010) are built by fully automati- cally analyzing translated text and learning the rules. We present two methods that merge ideas SMT has been embraced by the academic and com- from statistical machine translation (SMT) mercial research communities as the new dominant and translation memories (TM). We use a TM paradigm in machine translation. Almost all re- to retrieve matches for source segments, and cently published papers on machine translation are replace the mismatched parts with instructions to an SMT system to fill in the gap. We published on new SMT techniques. The method- show that for fuzzy matches of over 70%, one ology has left the research labs and become the method outperforms both SMT and TM base- basis of successful companies such as Language lines. Weaver and the highly visible Google and Microsoft web translation services. Even traditional rule-based companies such as Systran have embraced statisti- 1 Introduction cal methods and integrated them into their systems Two technological advances in the field of au- (Dugast et al., 2007). tomated language translation, translation memory The two technologies have not touched much in (TM) and statistical machine translation (SMT), the past not only because of the different devel- have seen vast progress over the last decades, but opment communities (software suppliers to transla- they have been developed very much in isolation.
    [Show full text]
  • Statistical Machine Translation
    ESSLLI Summer School 2008 Statistical Machine Translation Chris Callison-Burch, Johns Hopkins University Philipp Koehn, University of Edinburgh Intro to Statistical MT EuroMatrix MT Marathon Chris Callison-Burch Various approaches • Word-for-word translation • Syntactic transfer • Interlingual approaches • Controlled language • Example-based translation • Statistical translation 3 Advantages of SMT • Data driven • Language independent • No need for staff of linguists of language experts • Can prototype a new system quickly and at a very low cost Statistical machine translation • Find most probable English sentence given a foreign language sentence • Automatically align words and phrases within sentence pairs in a parallel corpus • Probabilities are determined automatically by training a statistical model using the parallel corpus 4 Parallel corpus what is more , the relevant cost im übrigen ist die diesbezügliche dynamic is completely under control. kostenentwicklung völlig unter kontrolle . sooner or later we will have to be früher oder später müssen wir die sufficiently progressive in terms of own notwendige progressivität der eigenmittel als resources as a basis for this fair tax grundlage dieses gerechten steuersystems system . zur sprache bringen . we plan to submit the first accession wir planen , die erste beitrittspartnerschaft partnership in the autumn of this year . im herbst dieses jahres vorzulegen . it is a question of equality and solidarity hier geht es um gleichberechtigung und . solidarität . the recommendation for the year 1999 die empfehlung für das jahr 1999 wurde vor has been formulated at a time of dem hintergrund günstiger entwicklungen favourable developments and optimistic und einer für den kurs der europäischen prospects for the european economy . wirtschaft positiven perspektive abgegeben .
    [Show full text]
  • Edinburgh's Statistical Machine Translation Systems for WMT16
    Edinburgh’s Statistical Machine Translation Systems for WMT16 Philip Williams1, Rico Sennrich1, Maria Nadejde˘ 1, Matthias Huck2, Barry Haddow1, Ondrejˇ Bojar3 1School of Informatics, University of Edinburgh 2Center for Information and Language Processing, LMU Munich 3Institute of Formal and Applied Linguistics, Charles University in Prague [email protected] [email protected] [email protected] [email protected] [email protected] [email protected] Abstract core syntax-based setup and experiments in Sec- tions 4 and 5. This paper describes the University of Ed- inburgh’s phrase-based and syntax-based 2 Phrase-based System Overview submissions to the shared translation tasks of the ACL 2016 First Conference on Ma- 2.1 Preprocessing chine Translation (WMT16). We sub- The training data was preprocessed us- mitted five phrase-based and five syntax- ing scripts from the Moses toolkit. based systems for the news task, plus one We first normalized the data using the phrase-based system for the biomedical normalize-punctuation.perl script, task. then performed tokenization (using the -a op- tion), and then truecasing. We did not perform 1 Introduction any corpus filtering other than the standard Moses method, which removes sentence pairs with Edinburgh’s submissions to the WMT 2016 news extreme length ratios, and sentences longer than translation task fall into two distinct groups: neu- 80 tokens. ral translation systems and statistical translation systems. In this paper, we describe the statisti- 2.2 Word Alignment cal systems, which includes a mix of phrase-based and syntax-based approaches.
    [Show full text]
  • Improved Learning for Machine Translation
    This document is part of the Research and Innovation Action “Quality Translation 21 (QT21)”. This project has received funding from the European Union’s Horizon 2020 program for ICT under grant agreement no. 645452. Deliverable D1.5 Improved Learning for Machine Translation Ondřej Bojar (CUNI), Jan-Thorsten Peter (RWTH), Weiyue Wang (RWTH), Tamer Alkhouli (RWTH), Yunsu Kim (RWTH), Miloš Stanojević (UvA), Khalil Sima’an (UvA), and Jon Dehdari (DFKI) Dissemination Level: Public 31st July, 2016 Quality Translation 21 D1.5: Improved Learning for Machine Translation Grant agreement no. 645452 Project acronym QT21 Project full title Quality Translation 21 Type of action Research and Innovation Action Coordinator Prof. Josef van Genabith (DFKI) Start date, duration 1st February, 2015, 36 months Dissemination level Public Contractual date of delivery 31st July, 2016 Actual date of delivery 31st July, 2016 Deliverable number D1.5 Deliverable title Improved Learning for Machine Translation Type Report Status and version Final (Version 1.0) Number of pages 78 Contributing partners CUNI, DFKI, RWTH, UVA WP leader CUNI Author(s) Ondřej Bojar (CUNI), Jan-Thorsten Peter (RWTH), Weiyue Wang (RWTH), Tamer Alkhouli (RWTH), Yunsu Kim (RWTH), Miloš Stanojević (UvA), Khalil Sima’an (UvA), and Jon Dehdari (DFKI) EC project officer Susan Fraser The partners in QT21 are: • Deutsches Forschungszentrum für Künstliche Intelligenz GmbH (DFKI), Germany • Rheinisch-Westfälische Technische Hochschule Aachen (RWTH), Germany • Universiteit van Amsterdam (UvA), Netherlands • Dublin City University (DCU), Ireland • University of Edinburgh (UEDIN), United Kingdom • Karlsruher Institut für Technologie (KIT), Germany • Centre National de la Recherche Scientifique (CNRS), France • Univerzita Karlova v Praze (CUNI), Czech Republic • Fondazione Bruno Kessler (FBK), Italy • University of Sheffield (USFD), United Kingdom • TAUS b.v.
    [Show full text]
  • A Survey of Statistical Machine Translation
    LAMP-TR-135 April 2007 CS-TR-4831 UMIACS-TR-2006-47 A Survey of Statistical Machine Translation Adam Lopez Computational Linguistics and Information Processing Laboratory Institute for Advanced Computer Studies Department of Computer Science University of Maryland College Park, MD 20742 [email protected] Abstract Statistical machine translation (SMT) treats the translation of natural language as a machine learning problem. By examining many samples of human-produced translation, SMT algorithms automatically learn how to translate. SMT has made tremendous strides in less than two decades, and many popular techniques have only emerged within the last few years. This survey presents a tutorial overview of state-of-the-art SMT at the beginning of 2007. We begin with the context of the current research, and then move to a formal problem description and an overview of the four main subproblems: translational equivalence modeling, mathematical modeling, parameter estimation, and decoding. Along the way, we present a taxonomy of some different approaches within these areas. We conclude with an overview of evaluation and notes on future directions. This is a revised draft of a paper currently under review. The contents may change in later drafts. Please send any comments, questions, or corrections to [email protected]. Feel free to cite as University of Maryland technical report UMIACS-TR-2006-47. The support of this research by the GALE program of the Defense Advanced Research Projects Agency, Contract No. HR0011-06-2-0001, ONR MURI Contract FCPO.810548265, and Department of Defense contract RD-02-5700 is acknowledged. Contents 1 Introduction 1 1.1 Background and Context .
    [Show full text]