Edinburgh's Statistical Machine Translation Systems for WMT16

Total Page:16

File Type:pdf, Size:1020Kb

Edinburgh's Statistical Machine Translation Systems for WMT16 Edinburgh’s Statistical Machine Translation Systems for WMT16 Philip Williams1, Rico Sennrich1, Maria Nadejde˘ 1, Matthias Huck2, Barry Haddow1, Ondrejˇ Bojar3 1School of Informatics, University of Edinburgh 2Center for Information and Language Processing, LMU Munich 3Institute of Formal and Applied Linguistics, Charles University in Prague [email protected] [email protected] [email protected] [email protected] [email protected] [email protected] Abstract core syntax-based setup and experiments in Sec- tions 4 and 5. This paper describes the University of Ed- inburgh’s phrase-based and syntax-based 2 Phrase-based System Overview submissions to the shared translation tasks of the ACL 2016 First Conference on Ma- 2.1 Preprocessing chine Translation (WMT16). We sub- The training data was preprocessed us- mitted five phrase-based and five syntax- ing scripts from the Moses toolkit. based systems for the news task, plus one We first normalized the data using the phrase-based system for the biomedical normalize-punctuation.perl script, task. then performed tokenization (using the -a op- tion), and then truecasing. We did not perform 1 Introduction any corpus filtering other than the standard Moses method, which removes sentence pairs with Edinburgh’s submissions to the WMT 2016 news extreme length ratios, and sentences longer than translation task fall into two distinct groups: neu- 80 tokens. ral translation systems and statistical translation systems. In this paper, we describe the statisti- 2.2 Word Alignment cal systems, which includes a mix of phrase-based and syntax-based approaches. We also include a For word alignment we used fast_align (Dyer et al., 2013)—except for German English, brief description of our phrase-based submission ↔ to the WMT16 biomedical translation task. Our where we used MGIZA++ (Gao and Vo- neural systems are described separately in Sen- gel, 2008)—followed by the standard nrich et al. (2016a). grow-diag-final-and symmetrization In most cases, our statistical systems build on heuristic. last year’s, incorporating recent modelling refine- 2.3 Language Models ments and adding this year’s new training data. For Romanian—a new language this year—we Our default approach to language modelling was paid particular attention to language-specific pro- to train individual models on each monolingual cessing of diacritics. For English Czech, we ex- corpus (except CommonCrawl) and then linearly- → perimented with a string-to-tree system, first using interpolate them to produce a single model. For Treex1 (formerly TectoMT; Popel and Žabokrt- some systems, we added separate neural or Com- ský, 2010) to produce Czech dependency parses, monCrawl LMs. Here we outline the various ap- then converting them to constituency representa- proaches and then in Section 3 we describe the tion and extracting GHKM rules. combination used for each language pair. In the next two sections, we describe the phrase- based systems, first describing the core setup in Interpolated LMs For individual monolingual Section 2 and then describing system-specific ex- corpora, we first used lmplz (Heafield et al., 2013) tensions and experimental results for each individ- to train count-based 5-gram language models with ual language pair in Section 3. We describe the modified Kneser-Ney smoothing (Chen and Good- man, 1998). We then used the SRILM toolkit 1http://ufal.mff.cuni.cz/treex (Stolcke, 2002) to linearly interpolate the models 399 Proceedings of the First Conference on Machine Translation, Volume 2: Shared Task Papers, pages 399–410, Berlin, Germany, August 11-12, 2016. c 2016 Association for Computational Linguistics using weights tuned to minimize perplexity on the 2.6 Decoding development set. In decoding we applied cube pruning (Huang and CommonCrawl LMs Our CommonCrawl lan- Chiang, 2007) with a stack size of 5000 (reduced guage models were trained in the same way as the to 1000 for tuning), Minimum Bayes Risk de- individual corpus-specific standard models, but coding (Kumar and Byrne, 2004), a maximum were not linearly-interpolated with other LMs. In- phrase length of 5, a distortion limit of 6, 100- stead, the log probabilities of CommonCrawl LMs best translation options and the no-reordering- were added as separate features of the systems’ over-punctuation heuristic (Koehn and Haddow, linear models. 2009). Neural LMs For some of our phrase-based sys- 3 Phrase-based Experiments tems we experimented with feed-forward neural 3.1 Finnish English network language models, both trained on target → n-grams only, and on “joint” or “bilingual” n- Similar to last year (Haddow et al., 2015), we built an unconstrained system for Finnish English us- grams (Devlin et al., 2014; Le et al., 2012). For → training these models we used the NPLM toolkit ing data extracted from OPUS (Tiedemann, 2012). (Vaswani et al., 2013), for which we have now im- Our parallel training set was the same as we used plemented gradient clipping to address numerical previously, but the language model training set issues often encountered during training. was extended with the addition of the news2015 monolingual corpus and the large WMT16 En- 2.4 Baseline Features glish CommonCrawl corpus. We used news- We follow the standard approach to SMT of scor- dev2015 for tuning, and newsdev2015 for testing ing translation hypotheses using a weighted lin- during system development. ear combination of features. The core features One clear problem that we noted with our sub- of our model are a 5-gram LM score (i.e. log mission from last year was the large number of probability), phrase translation and lexical trans- OOVs, which were then copied directly into the lation scores, word and phrase penalties, and a lin- English output. This is undoubtedly due to the ag- ear distortion score. The phrase translation prob- glutinative nature of Finnish, and probably was the abilities are smoothed with Good-Turing smooth- cause of our system being poorly judged by human ing (Foster et al., 2006). We used the hierarchi- evaluators, despite having a high BLEU score. To cal lexicalized reordering model (Galley and Man- address this, we split the Finnish input into sub- ning, 2008) with 4 possible orientations (mono- word units at both train and test time. In particular, tone, swap, discontinuous left and discontinuous we applied byte pair encoding (BPE) to split the right) in both left-to-right and right-to-left direc- Finnish source into smaller units, greatly reduc- tion. We also used the operation sequence model ing the vocabulary size. BPE is a technique which (OSM) (Durrani et al., 2013) with 4 count based has been recently used to good effect in neural ma- supportive features. We further employed domain chine translation (Sennrich et al., 2016b), where indicator features (marking which training cor- the models cannot handle large vocbaularies. It is pus each phrase pair was found in), binary phrase actually a merging algorithm, originally designed count indicator features, sparse phrase length fea- for compression, and works by starting with a tures, and sparse source word deletion, target word maximally split version of the training corpus (i.e. insertion, and word translation features (limited to split to characters) and iteratively merging com- the top K words in each language, typically with mon clusters. The merging continues for a speci- K = 50). fied number of iterations, and the merges are col- lected up to form the BPE model. At test time, 2.5 Tuning the recorded merges are applied to the test corpus, Since our feature set (generally around 500 to with the result that there are no OOVs in the test 1000 features) was too large for MERT, we used data. For the experiments here, we used 100,000 k-best batch MIRA for tuning (Cherry and Fos- BPE merges to create the model. ter, 2012). To speed up tuning we applied thresh- Applying BPE to Finnish English was clearly → old pruning to the phrase table, based on the direct effective at addressing the unknown word prob- translation model probability. lem, and in many cases the resulting translations 400 are quite understandable, e.g. into two parts randomly (so as to balance the “born source yös Intian on sanottu olevan kiinnostunut English” and “born Romanian” portions), using puolustusyhteistyösopimuksesta Japanin one for tuning and one for testing. For building the kanssa. final system, and for the contrastive experiments, base India is also said to be interested in puolus- we used the whole of newsdev2016 for tuning, and tusyhteistyösopimuksesta with Japan. newstest2016 for testing. bpe India is also said to be interested in defence In early experiments we noted that both the cooperation agreement with Japan. training and the development data were inconsis- reference India is also reportedly hoping for a tent in their use of diacritics leading to problems deal on defence collaboration between the with OOVs and sparse statistics. To address this two nations. we stripped off all diacritics from the Romanian However applying BPE to Finnish can also re- texts and the result was a significant increase in sult in some rather odd translations when it over- performance in our development setup. We also zealously splits: experimented with different language model com- source Balotelli oli vielä kaukana huippu- binations during development, with our submit- vireestään. ted system using three different language model base Balotelli was still far from huippuvireestään. features: a neural LM trained on just news2015 bpe Baloo, Hotel was still far from the peak of its monolingual, an n-gram language model trained vitality. on the WMT16 English CommonCrawl corpus, reference Balotelli is still far from his top tune. and a linear interpolation of language models We built four language models: an interpolated trained on all other WMT16 English corpora.
Recommended publications
  • (Meta-) Evaluation of Machine Translation
    (Meta-) Evaluation of Machine Translation Chris Callison-Burch Cameron Fordyce Philipp Koehn Johns Hopkins University CELCT University of Edinburgh ccb cs jhu edu fordyce celct it pkoehn inf ed ac uk Christof Monz Josh Schroeder Queen Mary, University of London University of Edinburgh christof dcs qmul ac uk j schroeder ed ac uk Abstract chine translation evaluation workshop. Here we ap- ply eleven different automatic evaluation metrics, This paper evaluates the translation quality and conduct three different types of manual evalu- of machine translation systems for 8 lan- ation. guage pairs: translating French, German, Beyond examining the quality of translations pro- Spanish, and Czech to English and back. duced by various systems, we were interested in ex- We carried out an extensive human evalua- amining the following questions about evaluation tion which allowed us not only to rank the methodologies: How consistent are people when different MT systems, but also to perform they judge translation quality? To what extent do higher-level analysis of the evaluation pro- they agree with other annotators? Can we im- cess. We measured timing and intra- and prove human evaluation? Which automatic evalu- inter-annotator agreement for three types of ation metrics correlate most strongly with human subjective evaluation. We measured the cor- judgments of translation quality? relation of automatic evaluation metrics with This paper is organized as follows: human judgments. This meta-evaluation re- veals surprising facts about the most com- • Section 2 gives an overview of the shared task. monly used methodologies. It describes the training and test data, reviews the baseline system, and lists the groups that 1 Introduction participated in the task.
    [Show full text]
  • Public Project Presentation and Updates
    7.2: Public Project Presentation and Updates Philipp Koehn Distribution: Public EuroMatrix Statistical and Hybrid Machine Translation Between All European Languages IST 034291 Deliverable 7.2 December 21, 2007 Project funded by the European Community under the Sixth Framework Programme for Research and Technological Development. Project ref no. IST-034291 Project acronym EuroMatrix Project full title Statistical and Hybrid Machine Translation Between All Eu- ropean Languages Instrument STREP Thematic Priority Information Society Technologies Start date / duration 01 September 2006 / 30 Months Distribution Public Contractual date of delivery September 1, 2007 Actual date of delivery September 13, 2007 Deliverable number 7.2 Deliverable title Public Project Presentation and Updates Type Report, contractual Status & version Finished Number of pages 19 Contributing WP(s) WP7 WP / Task responsible WP7 / 7.6 Other contributors none Author(s) Philipp Koehn EC project officer Xavier Gros Keywords The partners in EuroMatrix are: Saarland University (USAAR) University of Edinburgh (UEDIN) Charles University (CUNI-MFF) CELCT GROUP Technologies MorphoLogic For copies of reports, updates on project activities and other EuroMatrix-related in- formation, contact: The EuroMatrix Project Co-ordinator Prof. Hans Uszkoreit Universit¨at des Saarlandes, Computerlinguistik Postfach 15 11 50 66041 Saarbr¨ucken, Germany [email protected] Phone +49 (681) 302-4115- Fax +49 (681) 302-4700 Copies of reports and other material can also be accessed via the project’s homepage: http://www.euromatrix.net/ c 2007, The Individual Authors No part of this document may be reproduced or transmitted in any form, or by any means, electronic or mechanical, including photocopy, recording, or any information storage and retrieval system, without permission from the copyright owner.
    [Show full text]
  • 8.14: Annual Public Report
    8.14: Annual Public Report S. Busemann, O. Bojar, C. Callison-Burch, M. Cettolo, R. Garabik, J. van Genabith, P. Koehn, H. Schwenk, K. Simov, P. Wolf Distribution: Public EuroMatrixPlus Bringing Machine Translation for European Languages to the User ICT 231720 Deliverable 8.14 Project funded by the European Community under the Seventh Framework Programme for Research and Technological Development. Project ref no. ICT-231720 Project acronym EuroMatrixPlus Project full title Bringing Machine Translation for European Languages to the User Instrument STREP Thematic Priority ICT-2007.2.2 Cognitive systems, interaction, robotics Start date / duration 01 March 2009 / 38 Months Distribution Public Contractual date of delivery November 15, 2011 Actual date of delivery December 05, 2011 Date of last update December 5, 2011 Deliverable number 8.14 Deliverable title Annual Public Report Type Report Status & version Final Number of pages 13 Contributing WP(s) all WP / Task responsible DFKI Other contributors All Participants Author(s) S. Busemann, O. Bojar, C. Callison-Burch, M. Cettolo, R. Garabik, J. van Genabith, P. Koehn, H. Schwenk, K. Simov, P. Wolf EC project officer Michel Brochard Keywords The partners in DFKI GmbH, Saarbr¨ucken (DFKI) EuroMatrixPlus University of Edinburgh (UEDIN) are: Charles University (CUNI-MFF) Johns Hopkins University (JHU) Fondazione Bruno Kessler (FBK) Universit´edu Maine, Le Mans (LeMans) Dublin City University (DCU) Lucy Software and Services GmbH (Lucy) Central and Eastern European Translation, Prague (CEET) Ludovit Stur Institute of Linguistics, Slovak Academy of Sciences (LSIL) Institute of Information and Communication Technologies, Bulgarian Academy of Sciences (IICT-BAS) For copies of reports, updates on project activities and other EuroMatrixPlus-related information, contact: The EuroMatrixPlus Project Co-ordinator Prof.
    [Show full text]
  • Statistical and Hybrid Machine Translation Between All European Languages
    Statistical and hybrid machine translation between all European languages Publishable executive summary 01 September 2006 – 30 November 2007 Acronym: EuroMatrix Contract N°: IST 034291 Duration: 30 months from 01/09/2006 Coordinator: Prof. Dr. Hans Uszkoreit Saarland University D-66123 Saarbrücken [email protected] Scientific coordinator: Dr. Philipp Koehn Edinburgh University [email protected] PROJECT FUNDED BY THE EUROPEAN COMMUNITY UNDER THE SIXTH FRAMEWORK PROGRAMME FOR RESEARCH AND TECHNOLOGICAL DEVELOPMENT. Contractors involved Saarland University Germany Computational Linguistics The University of Edinburgh United Kingdom School of Informatics Univerzita Karlova v Praze/ Charles University Prague Czech Republic Institute of Formal and Applied Linguistics CELCT Center for the Evaluation of Language and Italy Communication Technologies MorphoLogic Hungary Germany Group Technologies AG -------------------------------------------------------------------------------------------- Publishable Executive Summary ©The EuroMatrix consortium, 2008 Summary description of project objectives EuroMatrix aims at a major push in machine translation technology applying the most advanced MT technologies systematically to all pairs of EU languages. Special attention is being paid to the languages of the new and near-term prospective member states. As part of this application development, EuroMatrix designs and investigates novel combinations of statistical techniques and linguistic knowledge sources as well as hybrid MT architectures. EuroMatrix addresses urgent European economic and social needs by concentrating on European languages and on high-quality translation to be employed for the publication of technical, social, legal and political documents. EuroMatrix aims at enriching the statistical MT approach with novel learning paradigms and experiment with new combinations of methods and resources from statistical MT, rule-based MT, shallow language processing and computational lexicography/morphology.
    [Show full text]
  • Abstract the Circle of Meaning: from Translation
    ABSTRACT Title of dissertation: THE CIRCLE OF MEANING: FROM TRANSLATION TO PARAPHRASING AND BACK Nitin Madnani, Doctor of Philosophy, 2010 Dissertation directed by: Professor Bonnie Dorr Department of Computer Science The preservation of meaning between inputs and outputs is perhaps the most ambitious and, often, the most elusive goal of systems that attempt to process natural language. Nowhere is this goal of more obvious importance than for the tasks of machine translation and paraphrase generation. Preserving meaning between the input and the output is paramount for both, the monolingual vs bilingual distinction notwithstanding. In this thesis, I present a novel, symbiotic relationship between these two tasks that I term the “circle of meaning”. Today’s statistical machine translation (SMT) systems require high quality human translations for parameter tuning, in addition to large bi-texts for learning the translation units. This parameter tuning usually involves generating translations at different points in the parameter space and obtaining feedback against human- authored reference translations as to how good the translations. This feedback then dictates what point in the parameter space should be explored next. To measure this feedback, it is generally considered wise to have multiple (usually 4) reference translations to avoid unfair penalization of translation hypotheses which could easily happen given the large number of ways in which a sentence can be translated from one language to another. However, this reliance on multiple reference translations creates a problem since they are labor intensive and expensive to obtain. Therefore, most current MT datasets only contain a single reference. This leads to the problem of reference sparsity—the primary open problem that I address in this dissertation— one that has a serious effect on the SMT parameter tuning process.
    [Show full text]
  • Improving Statistical Machine Translation Efficiency by Triangulation
    Improving Statistical Machine Translation Efficiency by Triangulation Yu Chen, Andreas Eisele, Martin Kay Saarland University Saarbrucken,¨ Germany {yuchen, eisele, kay}@coli.uni-sb.de Abstract In current phrase-based Statistical Machine Translation systems, more training data is generally better than less. However, a larger data set eventually introduces a larger model that enlarges the search space for the decoder, and consequently requires more time and more resources to translate. This paper describes an attempt to reduce the model size by filtering out the less probable entries based on testing correlation using additional training data in an intermediate third language. The central idea behind the approach is triangulation, the process of incorporating multilingual knowledge in a single system, which eventually utilizes parallel corpora available in more than two languages. We conducted experiments using Europarl corpus to evaluate our approach. The reduction of the model size can be up to 70% while the translation quality is being preserved. 1. Introduction 2. Phrase-based SMT The dominant paradigm in SMT is referred to as phrase- Statistical machine translation (SMT) is now generally based machine translation. The term refers to a sequence of taken to be an enterprise in which machine learning tech- words characterized by its statistical rather then any gram- niques are applied to a bilingual corpus to produce a trans- matical properties. At the center of the process is a phrase lation system entirely automatically. Such a scheme has table, a list of phrases identified in a source sentence to- many potential advantages over earlier systems which re- gether with potential translations.
    [Show full text]
  • Factored Translation Models and Discriminative Training For
    Machine Translation at Edinburgh Factored Translation Models and Discriminative Training Philipp Koehn, University of Edinburgh 9 July 2007 Philipp Koehn, University of Edinburgh EuroMatrix 9 July 2007 1 Overview • Intro: Machine Translation at Edinburgh • Factored Translation Models • Discriminative Training Philipp Koehn, University of Edinburgh EuroMatrix 9 July 2007 2 The European Challenge Many languages • 11 official languages in EU-15 • 20 official languages in EU-25 • many more minority languages Challenge • European reports, meetings, laws, etc. • develop technology to enable use of local languages as much as possible Philipp Koehn, University of Edinburgh EuroMatrix 9 July 2007 3 Existing MT systems for EU languages [from Hutchins, 2005] Cze Dan Dut Eng Est Fin Fre Ger Gre Hun Ita Lat Lit Mal Pol Por Slo Slo Spa Swe Czech –..1..11..1.........4 Danish .–.....1............1 Dutch ..–6..21............9 English 2 . 6 – . 42 48 3 3 29 1 . 7 30 2 . 48 1 222 Estonian ....–...............0 Finnish ...2.–.1............3 French 1 . 2 38 . – 22 3 . 9 . 1 5 . 10 . 91 German 1 1 1 49 . 1 23 – . 1 8 . 4 3 2 . 8 1 103 Greek ...2..3.–...........5 Hungarian ...1...1.–..........2 Italian 1 . 25 . 9 8 . – . 1 3 . 7 . 54 Latvian ...1.......–........1 Lithuanian ............–.......0 Maltese .............–......0 Polish ...6..13..1...–2..1.14 Portuguese . 25 . 4 4 . 3 . 1 – . 6 . 43 Slovak ...1...1........–...2 Slovene .................–..0 Spanish 1 . 42 . 8 7 . 7 . 1 6 . – . 72 Swedish ...2...1...........–3 6 1 9 201 0 1 93 99 6 4 58 1 0 0 15 49 4
    [Show full text]
  • Convergence of Translation Memory and Statistical Machine Translation
    Convergence of Translation Memory and Statistical Machine Translation Philipp Koehn Jean Senellart University of Edinburgh Systran 10 Crichton Street La Grande Arche Edinburgh, EH8 9AB 1, Parvis de la Defense´ Scotland, United Kingdom 92044 Paris, France [email protected] [email protected] Abstract tems (Koehn, 2010) are built by fully automati- cally analyzing translated text and learning the rules. We present two methods that merge ideas SMT has been embraced by the academic and com- from statistical machine translation (SMT) mercial research communities as the new dominant and translation memories (TM). We use a TM paradigm in machine translation. Almost all re- to retrieve matches for source segments, and cently published papers on machine translation are replace the mismatched parts with instructions to an SMT system to fill in the gap. We published on new SMT techniques. The method- show that for fuzzy matches of over 70%, one ology has left the research labs and become the method outperforms both SMT and TM base- basis of successful companies such as Language lines. Weaver and the highly visible Google and Microsoft web translation services. Even traditional rule-based companies such as Systran have embraced statisti- 1 Introduction cal methods and integrated them into their systems Two technological advances in the field of au- (Dugast et al., 2007). tomated language translation, translation memory The two technologies have not touched much in (TM) and statistical machine translation (SMT), the past not only because of the different devel- have seen vast progress over the last decades, but opment communities (software suppliers to transla- they have been developed very much in isolation.
    [Show full text]
  • Statistical Machine Translation
    ESSLLI Summer School 2008 Statistical Machine Translation Chris Callison-Burch, Johns Hopkins University Philipp Koehn, University of Edinburgh Intro to Statistical MT EuroMatrix MT Marathon Chris Callison-Burch Various approaches • Word-for-word translation • Syntactic transfer • Interlingual approaches • Controlled language • Example-based translation • Statistical translation 3 Advantages of SMT • Data driven • Language independent • No need for staff of linguists of language experts • Can prototype a new system quickly and at a very low cost Statistical machine translation • Find most probable English sentence given a foreign language sentence • Automatically align words and phrases within sentence pairs in a parallel corpus • Probabilities are determined automatically by training a statistical model using the parallel corpus 4 Parallel corpus what is more , the relevant cost im übrigen ist die diesbezügliche dynamic is completely under control. kostenentwicklung völlig unter kontrolle . sooner or later we will have to be früher oder später müssen wir die sufficiently progressive in terms of own notwendige progressivität der eigenmittel als resources as a basis for this fair tax grundlage dieses gerechten steuersystems system . zur sprache bringen . we plan to submit the first accession wir planen , die erste beitrittspartnerschaft partnership in the autumn of this year . im herbst dieses jahres vorzulegen . it is a question of equality and solidarity hier geht es um gleichberechtigung und . solidarität . the recommendation for the year 1999 die empfehlung für das jahr 1999 wurde vor has been formulated at a time of dem hintergrund günstiger entwicklungen favourable developments and optimistic und einer für den kurs der europäischen prospects for the european economy . wirtschaft positiven perspektive abgegeben .
    [Show full text]
  • Improved Learning for Machine Translation
    This document is part of the Research and Innovation Action “Quality Translation 21 (QT21)”. This project has received funding from the European Union’s Horizon 2020 program for ICT under grant agreement no. 645452. Deliverable D1.5 Improved Learning for Machine Translation Ondřej Bojar (CUNI), Jan-Thorsten Peter (RWTH), Weiyue Wang (RWTH), Tamer Alkhouli (RWTH), Yunsu Kim (RWTH), Miloš Stanojević (UvA), Khalil Sima’an (UvA), and Jon Dehdari (DFKI) Dissemination Level: Public 31st July, 2016 Quality Translation 21 D1.5: Improved Learning for Machine Translation Grant agreement no. 645452 Project acronym QT21 Project full title Quality Translation 21 Type of action Research and Innovation Action Coordinator Prof. Josef van Genabith (DFKI) Start date, duration 1st February, 2015, 36 months Dissemination level Public Contractual date of delivery 31st July, 2016 Actual date of delivery 31st July, 2016 Deliverable number D1.5 Deliverable title Improved Learning for Machine Translation Type Report Status and version Final (Version 1.0) Number of pages 78 Contributing partners CUNI, DFKI, RWTH, UVA WP leader CUNI Author(s) Ondřej Bojar (CUNI), Jan-Thorsten Peter (RWTH), Weiyue Wang (RWTH), Tamer Alkhouli (RWTH), Yunsu Kim (RWTH), Miloš Stanojević (UvA), Khalil Sima’an (UvA), and Jon Dehdari (DFKI) EC project officer Susan Fraser The partners in QT21 are: • Deutsches Forschungszentrum für Künstliche Intelligenz GmbH (DFKI), Germany • Rheinisch-Westfälische Technische Hochschule Aachen (RWTH), Germany • Universiteit van Amsterdam (UvA), Netherlands • Dublin City University (DCU), Ireland • University of Edinburgh (UEDIN), United Kingdom • Karlsruher Institut für Technologie (KIT), Germany • Centre National de la Recherche Scientifique (CNRS), France • Univerzita Karlova v Praze (CUNI), Czech Republic • Fondazione Bruno Kessler (FBK), Italy • University of Sheffield (USFD), United Kingdom • TAUS b.v.
    [Show full text]
  • A Survey of Statistical Machine Translation
    LAMP-TR-135 April 2007 CS-TR-4831 UMIACS-TR-2006-47 A Survey of Statistical Machine Translation Adam Lopez Computational Linguistics and Information Processing Laboratory Institute for Advanced Computer Studies Department of Computer Science University of Maryland College Park, MD 20742 [email protected] Abstract Statistical machine translation (SMT) treats the translation of natural language as a machine learning problem. By examining many samples of human-produced translation, SMT algorithms automatically learn how to translate. SMT has made tremendous strides in less than two decades, and many popular techniques have only emerged within the last few years. This survey presents a tutorial overview of state-of-the-art SMT at the beginning of 2007. We begin with the context of the current research, and then move to a formal problem description and an overview of the four main subproblems: translational equivalence modeling, mathematical modeling, parameter estimation, and decoding. Along the way, we present a taxonomy of some different approaches within these areas. We conclude with an overview of evaluation and notes on future directions. This is a revised draft of a paper currently under review. The contents may change in later drafts. Please send any comments, questions, or corrections to [email protected]. Feel free to cite as University of Maryland technical report UMIACS-TR-2006-47. The support of this research by the GALE program of the Defense Advanced Research Projects Agency, Contract No. HR0011-06-2-0001, ONR MURI Contract FCPO.810548265, and Department of Defense contract RD-02-5700 is acknowledged. Contents 1 Introduction 1 1.1 Background and Context .
    [Show full text]
  • Statistical Machine Translation: the Basic, the Novel, and the Speculative
    Statistical Machine Translation: the basic, the novel, and the speculative Philipp Koehn, University of Edinburgh 4 April 2006 Philipp Koehn SMT Tutorial 4 April 2006 1 The Basic • Translating with data – how can computers learn from translated text? – what translated material is out there? – is it enough? how much is needed? • Statistical modeling – framing translation as a generative statistical process • EM Training – how do we automatically discover hidden data? • Decoding – algorithm for translation Philipp Koehn SMT Tutorial 4 April 2006 2 The Novel • Automatic evaluation methods – can computers decide what are good translations? • Phrase-based models – what are atomic units of translation? – the best method in statistical machine translation • Discriminative training – what are the methods that directly optimize translation performance? Philipp Koehn SMT Tutorial 4 April 2006 3 The Speculative • Syntax-based transfer models – how can we build models that take advantage of syntax? – how can we ensure that the output is grammatical? • Factored translation models – how can we integrate different levels of abstraction? Philipp Koehn SMT Tutorial 4 April 2006 4 The Rosetta Stone • Egyptian language was a mystery for centuries • 1799 a stone with Egyptian text and its translation into Greek was found ⇒ Humans could learn how to translated Egyptian Philipp Koehn SMT Tutorial 4 April 2006 5 Parallel Data • Lots of translated text available: 100s of million words of translated text for some language pairs – a book has a few 100,000s words
    [Show full text]