A Rule-Based Shallow-Transfer Machine Translation System for Scots and English

Total Page:16

File Type:pdf, Size:1020Kb

A Rule-Based Shallow-Transfer Machine Translation System for Scots and English A Rule-based Shallow-transfer Machine Translation System for Scots and English Gavin Abercrombie [email protected] Abstract An open-source rule-based machine translation system is developed for Scots, a low-resourced minor language closely related to English and spoken in Scotland and Ireland. By concentrating on translation for assimilation (gist comprehension) from Scots to English, it is proposed that the development of dictionaries designed to be used within the Apertium platform will be sufficient to produce translations that improve non-Scots speakers understanding of the language. Mono- and bilingual Scots dictionaries are constructed using lexical items gathered from a variety of resources across several domains. Although the primary goal of this project is translation for gisting, the system is evaluated for both assimilation and dissemination (publication-ready translations). A variety of evaluation methods are used, including a cloze test undertaken by human volunteers. While evaluation results are comparable to, and in some cases superior to, those of other language pairs within the Apertium platform, room for improvement is identified in several areas of the system. Keywords: Scots, machine translation, assimilation 1. Introduction (‘what did he do that for’). There are also notable differ- The Scots language is spoken by over 1.5 million people ences to English in negation e.g. huvnae (‘have not’), din- in Scotland1 and a further 140,000 in Ireland,2 but is low- nae (‘do not’), and the double modal e.g. he’ll can dae that resourced in terms of technology - as far as I am aware there (‘he’ll be able to do that’) (Beal, 1997). are currently no automatic translators available for this lan- While Scots shares a great deal of vocabulary with English, guage. The development of machine translation (MT) sys- albeit often with different forms and/or senses, it also in- tems can lead to positive outcomes for a minor language cludes a large quantity of words derived from Old English and its speakers by contributing to the language’s standard- that are not present in English, as well as loan words from isation, and assisting the diffusion of content originally cre- Gaelic, French and Latin, the Scandinavian languages, and ated in that language (Forcada, 2006). The aim of this other West Germanic languages such as Frisian (Jones, project is to take a step towards the ‘de-minorizing’ of Scots 1997). 4 by developing a shallow transfer, rule-based MT system as There is no official standard written form of Scots, so part of the Apertium platform. spelling varies greatly - any translation system needs to ac- Since all Scots speakers comprehend standard English,3 count for multiple forms of many, if not most, words. this study will focus on translation from Scots to English 2.2. Translation: assimilation and dissemination for assimilation (or gisting) purposes. With assimilation Dis- thought to be the most common use of MT systems today The aims of translation between languages can vary. semination (O’Regan et al., 2013), this would seem to be an effective is ‘the production of translations of ‘publishable use of time and resources as an initial goal in MT between quality’ (Hutchins, 2003a), and can help in the creation of 5 assimila- these two languages. As the two languages are very closely more text in a lesser-resourced language, while tion related, it is proposed that translation of the bulk of lexical is translation for gist comprehension by readers who items in Scots texts (combined with part-of-speech (POS) ‘can accept poor quality as long as they can get an idea of analysis) will be sufficent to enable gist comprehension by what the text conveys’ (Hutchins, 2003b). non-Scots speakers. Successful assimilation is often viewed as the more realis- tic and achievable goal for MT systems (Hutchins, 2003a) 2. Background and is probably ‘the most frequent application of MT nowa- 2.1. The Scots language days’ (O’Regan et al., 2013). Translation for assimilation would enable people who do not speak the language to Scots is a West Germanic language descended from Old understand Scots text ‘thus removing an argument against Northumbrian (a branch of Old English) and closely related writing in the lesser-resourced language,’6 and is the pri- to modern English (Macafee and Aitken, 2002). Indeed, it mary aim of this study. shares many characteristics with its neighbour and relative including SVO sentence structure and a ‘vast body’ of com- 2.3. Machine translation mon vocabulary (Tulloch, 1997). Approaches to machine translation can be broadly catego- Grammatical differences can however be seen in the word rized as either rule-based or statistical, or in some cases, a order of some constructions e.g. whit fur did he dae that? hybrid of the two. 1 Scotland’s Census 2011 - National Records of Scotland 4www.educationscotland.gov.uk/knowledgeoflanguage 2 2001 Census of Northern Ireland - Northern Ireland /scots/writinginscots/scotsspelling/index.asp Statistics & Research Agency 5wiki.apertium.org/wiki/Assimilation and Dissemination 3 Scotland’s Census 2001 - National Records of Scotland 6wiki.apertium.org/wiki/Assimilation and Dissemination 578 While statistical MT systems have the advantage of not re- The modules perform the following functions: quiring the explicit programming of language rules, they • The deformatter and reformatter modules handle for- generally require large amounts of bilingual corpus data in matting and blank spaces. order to produce satisfactory results. As Forcada (2006) suggests, lacking such resources, ‘it may be much easier • The morphological analyser maps each word (or for speakers of the minor language to encode the language multi-word unit) in the input to entries in a monolin- expertise needed to build a rule-based machine translation gual dictionary, returning their lemma, lexical cate- system.’ gory and morphological information. This informa- While statistical MT tends to produce more fluent transla- tion is passed to the other modules and eventually tions, those of rule-based systems are often more faithful mapped to a target language surface form by the mor- to the details of content of the original text (Forcada et al., phological generator. 2011), a somewhat important advantage where information • A constraint grammar module contains rules designed transferral is concerned. to reduce ambiguity regarding parts-of-speech. Due to the low-resource status of Scots, and this study’s focus on translation for assimilation, rule-based translation • The part-of-speech tagger selects the most statistically would currently seem to be the more appropriate approach. likely of the possible lexical forms, given an item’s context. 2.3.1. Shallow-transfer rule-based machine • The lexical transfer module maps the given lexical translation item to its corrsponding lemma in the bilingual dic- Most rule-based MT systems create translations by parsing tionary. the source text, creating an intermediate symbolic represen- • tation of it, and then generating a final translation in the tar- The chunker segments the input into syntactic chunks get language. They apply mappings between lexical items and handles transfer rules. stored in dictionaries as well as transfer rules to account for • The post generator handles orthographic operations structural differences between the two languages. Trans- such as contractions. lation of unrelated or distantly related languages requires deep syntactic and semantic analysis, whereas closely re- Further details of the Apertium architecture can be found lated languages (such as English and Scots) can be trans- in Forcada et al. (2011). lated with shallow parsing (Sanchez-Mart´ ´ınez and Forcada, 3. Development of the English-Scots 2009). translation system In the belief that translation of Scots-only lexical items to 2.3.2. The Apertium MT platform English will be sufficient to enable gist comprehension, this is a free, open-source platform for develop- Apertium study concentrates on the development of the monolingual ing rule-based, shallow-transfer machine translation sys- and bilingual dictionaries for this language pair, and partic- tems, which was initially developed for translation between ularly on including Scots words that do not exist in English. closely related languages (Forcada et al., 2011). As ‘one of Development of the system was undertaken in the following the few open-source MT systems that can be used for real- stages. life purposes’ (Forcada, 2006), the Apertium platform is highly suitable for this project. 3.1. Prior status of the eng-sco language pair in The system consists of a number of modules through which the Apertium platform the input text is passed and modified before a translation in Apertium language pairs are classified as being in one the target language is outputted. Figure 1 illustrates this of four stages of development: from the trunk for release pipeline. quality pairs, down through staging and nursery to incuba- tor where incomplete ‘dictionaries, dictionary fragments, rules, things that aren’t quite ready to live in the real world’ are kept.7 Prior to this study, English & Scots (eng-sco) was in this latter category, and contained a Scots dictionary of only 172 entries. 3.2. Construction of eng-sco translator modules An English monolingual dictionary containing ap- proximately 20,000 lexical entries and correspond- ing morphological information was first added to eng-sco in the Apertium incubator. Associated dictionary processing files such as the post generator apertium-eng.post-eng.dix were also included. Figure 1: Architecture of the Apertium system. Adapted Work then began on expanding the Scots dictionaries. from Forcada et al., (2011). 7http://wiki.apertium.org/wiki/Incubator 579 3.3. Acquisition of linguistic data ↓ Scots lexical items were collected from the following Morphological analysis sources and added to the monolingual Scots dictionary, the English-Scots bilingual dictionary, and where necessary the ∧Ye/Prpers < prn >< subj >< p2 ><mf ><sp> monolingual English dictionary.
Recommended publications
  • "Machine Translation Evaluation Through Post-Editing…"
    ©inTRAlinea & Anna Fernández Torné (2016). "Machine translation evaluation through post-editing measures in audio description", inTRAlinea Vol. 18. Permanent URL: http://www.intralinea.org/archive/article/2200 inTRAlinea [ISSN 1827-000X] is the online translation journal of the Department of Interpreting and Translation (DIT) of the University of Bologna, Italy. This printout was generated directly from the online version of this article and can be freely distributed under the following Creative Commons License. Machine translation evaluation through post-editing measures in audio description By Anna Fernández Torné (Universitat Autònoma de Barcelona, Spain) Abstract & Keywords English: The number of accessible audiovisual products and the pace at which audiovisual content is made accessible need to be increased, reducing costs whenever possible. The implementation of different technologies which are already available in the translation field, specifically machine translation technologies, could help reach this goal in audio description for the blind and partially sighted. Measuring machine translation quality is essential when selecting the most appropriate machine translation engine to be implemented in the audio description field for the English-Catalan language combination. Automatic metrics and human assessments are often used for this purpose in any specific domain and language pair. This article proposes a methodology based on both objective and subjective measures for the evaluation of five different and free online machine translation systems. Their raw machine translation outputs and the post-editing effort that is involved are assessed using eight different scores. Results show that there are clear quality differences among the systems assessed and that one of them is the best rated in six out of the eight evaluation measures used.
    [Show full text]
  • A Finite-State Morphological Analyser for Sindhi
    A Finite-State Morphological Analyser for Sindhi Raveesh Motlani1, Francis M. Tyers2 and Dipti M. Sharma1 1FC Kohli Center on Intelligent Systems (KCIS), International Institute of Information Technology Hyderabad, Telangana, India 2 HSL-fakultehta, UiT Norgga árktalaš universitehta N-9019 Romsa [email protected], [email protected], [email protected] Abstract Morphological analysis is a fundamental task in natural-language processing, which is used in other NLP applications such as part-of-speech tagging, syntactic parsing, information retrieval, machine translation, etc. In this paper, we present our work on the development of free/open-source finite-state morphological analyser for Sindhi. We have used Apertium’s lttoolbox as our finite-state toolkit to implement the transducer. The system is developed using a paradigm-based approach, wherein a paradigm defines all the word forms and their morphological features for a given stem (lemma). We have evaluated our system on the Sindhi Wikipedia, which is a freely-available large corpus of Sindhi and achieved a reasonable coverage of about 81% and a precision of over 97%. Keywords: Sindhi, Morphological Analysis, Finite-State Machines [ɓ] ٻ [ɲ] ڃ [ŋ] ڱ Introduction .1 [ɠ] [ʄ] [bʱ] ڀ ڄ ڳ Morphology describes the internal structure of words in a [dʱ] ڌ [cʰ] ڇ [k] ڪ language. A morphological analysis of a word involves [ɗ] ڏ [ʈʰ] ٺ [ɳ] ڻ ,describing one or more of its properties such as: gender [ɖ] ڊ [ʈ] ٽ [pʰ] ڦ number, person, case, lexical category, etc. Morphological [ɖʱ] ڍ [tʰ] ٿ [ɽ] ڙ -analysis of a word thus becomes a fundamental and cru cial task in natural-language processing for any language.
    [Show full text]
  • Multi-Script Morphological Transducers and Transcribers for Seven Turkic
    Multi-script morphological transducers and transcribers for seven Turkic languages Jonathan Washington, Francis Tyers, Oğuzhan Kuyrukçu Swarthmore College, Indiana University / Высшая Школа Экономики, Boğaziçi Üniversitesi [email protected], [email protected], [email protected] This paper describes ongoing work to augment morphological transducers for seven Turkic languages with support for multiple scripts each, as well as respective IPA transcription systems. Evaluation demonstrates that our approach yields coverage equivalent to or not much lower than that of the base transducers. Background. A morphological transducer converts between form and analysis, e.g. алмалардан ↔ алма<n><pl> <abl>, where the form-to-analysis task is termed “morphological analysis” and the analysis-to-form task is termed “morphological generation”. Existing Free/Open-Source morphological transducers for Turkic languages (Wash- ington et al. 2020) are implemented in only one orthography, despite a number of the languages being currently written in two or more orthographies, or having a large body of text written in an orthography that was recently switched away from. This paper builds on work in which Cyrillic support was added to a transducer for Crimean Tatar which had been implemented in the Latin script (Tyers et al. 2019). We leverage morphological transducers for Kazakh (imple- mented in the Cyrillic script), Kyrgyz (Cyrillic), Turkmen (Latin), Qaraqalpaq (Latin), Uzbek (Latin), and Uyghur (Perso-Arabic), and add support for analysis and generation in additional scripts that are currently or have recently been used for the languages. Specifically, we add Cyrillic support to Turkmen, Qaraqalpaq, Uzbek, and Uyghur transducers; Perso-Arabic support to the Kazakh and Kyrgyz transducers; Latin script to the Kazakh and Uyghur transducers; and IPA support to all of them.
    [Show full text]
  • Developing Morph-Analyzer for Urdu Using Apertium: Some Issues
    Developing Morph-Analyzer for Urdu Using Apertium: Some issues Shahid Mushataq Bhat [email protected] Linguistic Data Consortium for Indian Languages (LDCIL) CIIL, Mysore Content: Introduction Morphological features of Urdu Apertium (LT-toolbox): Some background Computing Noun-morphology using LT Toolbox Split-Orthography of Urdu Conclusion Introduction: Automatic morphological analysis is the fundamental task in NLP that can be employed in enhancing the accuracy of POS-taggers, Chunkers, Parsers and Information retrieval systems. Computational morphology models the internal structure of words i-e the way; words are built out of minimal units called morphemes. Most of natural languages construct words by concatenating morphemes together in strict orders. Such Concatenative morphotactics is highly productive, particularly, in agglutinative languages like Tamil, Kannada, Manipuri, etc but in some languages like Hebrew and Arabic (Semitic languages) infixation is the main morphological operation (instead of concatenation), constituting Non-Concatenative (Templatic or Root & Pattern) morphology. Continues …… Beyond this Concatenative and Non-Concatenative polarity, Urdu nouns (unlike nouns of other Indian Languages) show the interplay of both types of morphologies. So, morphological structure of Urdu like Tagalog (a language of Philippines) can’t be computed adequately unless dual nature of its morphology is not taken into account. “The morphotactic limitations of the traditional implementations are the direct result of relying solely
    [Show full text]
  • Estudio Comparativo De Tres
    TRABAJO DE FIN DE GRADO FACULTAD DE CIENCIAS HUMANAS Y SOCIALES GRADO EN TRADUCCIÓN E INTERPRETACIÓN ESTUDIO COMPARATIVO DE TRES TRADUCTORES AUTOMÁTICOS EN LÍNEA : DEEP L, YANDEX Y APERTIUM Autora: Mónica Adán Soriano Directora: Profesora Mª Luisa Romana García Madrid, junio 2019 Resumen : Este trabajo tiene la finalidad de comparar traductores automáticos en línea para así determinar cuál es el traductor más avanzado para un texto técnico. Para ello, primero, habrá un análisis de la evolución histórica de la traducción automática y sus usos, así como los programas desarrollados para ello. Se explicarán además los diferentes sistemas de traducción automática que existen y cómo funcionan. En la parte experimental, se escogerá un texto de carácter técnico en español y de oraciones que supongan un reto para un traductor y se procesará en los tres traductores automáticos escogidos según su modalidad para realizar una traducción al inglés. Una vez obtenidas las respuestas, se analizarán los errores cometidos por los traductores automáticos, concluyendo así con los errores más comunes y el mejor traductor automático en línea así como los usos que se le pueden dar. Palabras clave: traducción automática, sistemas basados en estadística, Yandex, sistemas neuronales, DeepL, sistemas basados en reglas, Apertium, errores de traducción. Abstract : The aim of this dissertation is to do a comparative research analysis on three automatic translation programs available on the internet to determine which is the most advanced for a technical text. First, we will do an analysis on the historic evolution of automatic translation and the several uses given to it, as well as the diverse programs developed for it and how they work.
    [Show full text]
  • The Apertium Bilingual Dictionaries on the Web of Data
    Undefined 0 (0) 1 1 IOS Press The Apertium Bilingual Dictionaries on the Web of Data Jorge Gracia a;∗, Marta Villegas b, Asunción Gómez-Pérez a, and Núria Bel b, a Ontology Engineering Group, Universidad Politécnica de Madrid Campus de Montegancedo s/n Boadilla del Monte 28660 Madrid. Spain E-mail: {jgracia,asun}@fi.upm.es b Institut Universitari Linguistica Aplicada, Universitat Pompeu Fabra Roc Boronat, 138 08018 Barcelona. Spain E-mail: {marta.villegas,nuria.bel}@upf.edu Abstract. Bilingual electronic dictionaries contain collections of lexical entries in two languages, with explicitly declared trans- lation relations between such entries. Nevertheless, they are typically developed in isolation, in their own formats and accessible through proprietary APIs. In this paper we propose the use of Semantic Web techniques to make translations available on the Web to be consumed by other semantic enabled resources in a direct manner, based on standard languages and query means. In particular, we describe the conversion of the Apertium family of bilingual dictionaries and lexicons into RDF (Resource Descrip- tion Framework) and how their data have been made accessible on the Web as linked data. As result, all the converted dictionaries (many of them covering under-resourced languages) are connected among them and can be easily traversed from one to another to obtain, for instance, translations between language pairs not originally connected in any of the original dictionaries. Keywords: linguistic linked data, multilingualism, Apertium, bilingual dictionaries, lexicons, lemon, translation 1. Introduction In this article we will focus on the case of electronic bilingual dictionaries as a particular type of languages The publication of bilingual and multilingual lan- resources.
    [Show full text]
  • Apertium: a Free/Open-Source Platform for Machine Translation and Basic Language Technology Mikel L
    Apertium: A free/open-source platform for machine translation and basic language technology www.apertium.org Mikel L. Forcada1 Francis M. Tyers2,3 1Departament de Llenguatges i Sistemes Inform`atics, Universitat d’Alacant, E-03690 Sant Vicent del Raspeig (Spain) 2Department of Linguistics, Indiana University, IN 47405, U.S.A. 3 School of Linguistics, Higher School of Economics, Moscow, Russia [email protected], [email protected] Apertium components Languages and language pairs Since 2005, Apertium provides: 1. A free/open-source, modular, shallow/deep transfer, language-independent Language data is encoded mostly in XML, but some language pairs contain data machine translation engine with: text format management, finite-state lexical encoded in other text-based formats. processing, statistical and constraint-based lexical disambiguation, Stable language pairs (bilingual data) include: discontiguous multi-word assembly and disassembly, anaphora resolution, and shallow structural transfer based on finite-state pattern matching. afr hbs bel crh kaz ind mlt pol urd 2. Free/open-source linguistic data in well-specified XML formats for a variety of languages and language pairs nld bre epo eus cym mkd slv isl rus tur tat zsm ara szl hin 3. Free/open-source tools: compilers to turn linguistic data into an efficient fra oci eng bul swe ukr form used by the engine and software to learn disambiguation or translation spa dan rules from corpora. Interfaces with external technologies: HFST (for some morphologically-rich arg ast por ron sme nno languages), VISL CG-3 (for rule-based lexical disambiguation). cat glg nob Machine translation but not only! ita srd For many of these languages, there are separate monolingual packages.
    [Show full text]
  • Apertium-Icenlp: a Rule-Based Icelandic to English Machine Translation System
    Apertium-IceNLP: A rule-based Icelandic to English machine translation system Martha Dís Brandt, Hrafn Loftsson, Francis M. Tyers Hlynur Sigurþórsson School of Computer Science Dept. Lleng. i. Sist. Inform. Reykjavik University Universitat d’Alacant IS-101 Reykjavik, Iceland E-03071 Alacant, Spain {marthab08,hrafn,hlynurs06}@ru.is [email protected] Abstract translate between the source language (SL) and the target language (TL). We describe the development of a pro- SMT has many advantages, e.g. it is data-driven, totype of an open source rule-based language independent, does not need linguistic ex- Icelandic English MT system, based on perts, and prototypes of new systems can by built → the Apertium MT framework and IceNLP, quickly and at a low cost. On the other hand, the a natural language processing toolkit for need for parallel corpora as training data in SMT Icelandic. Our system, Apertium-IceNLP, is also its main disadvantage, because such corpora is the first system in which the whole are not available for a myriad of languages, espe- morphological and tagging component of cially the so-called less-resourced languages, i.e. Apertium is replaced by modules from languages for which few, if any, natural language an external system. Evaluation shows processing (NLP) resources are available. When that the word error rate and the position- there is a lack of parallel corpora, other machine independent word error rate for our pro- translation (MT) methods, such as rule-based MT, totype is 50.6% and 40.8%, respectively. e.g. Apertium (Forcada et al., 2009), may be used As expected, this is higher than the corre- to create MT systems.
    [Show full text]
  • Proceedings of the 21St Annual Conference of the European Association for Machine Translation
    Proceedings of the 21st Annual Conference of the European Association for Machine Translation 28{30 May 2018 Universitat d'Alacant Alacant, Spain Edited by Juan Antonio P´erez-Ortiz Felipe S´anchez-Mart´ınez Miquel Espl`a-Gomis Maja Popovi´c Celia Rico Andr´eMartins Joachim Van den Bogaert Mikel L. Forcada Organised by research group The papers published in this proceedings are |unless indicated otherwise| covered by the Creative Commons Attribution-NonCommercial-NoDerivatives 3.0 International (CC-BY-ND 3.0). You may copy, distribute, and transmit the work, provided that you attribute it (au- thorship, proceedings, publisher) in the manner specified by the author(s) or licensor(s), and that you do not use it for commercial purposes. The full text of the licence may be found at https://creativecommons.org/licenses/by-nc-nd/3.0/deed.en. c 2018 The authors ISBN: 978-84-09-01901-4 Rule-based machine translation from Kazakh to Turkish Sevilay Bayatli Sefer Kurnaz Dept. Elec. and Computer Engineering Dept. Elec. and Computer Engineering Altınbaş Üniversitesi Altınbaş Üniversitesi [email protected] [email protected] Ilnar Salimzianov Jonathan North Washington Francis M. Tyers School of Science and Technology Linguistics Deptartment School of Linguistics Nazarbayev University Swarthmore College Higher School of Economics [email protected] [email protected] [email protected] Abstract all the dictionary entries and transfer rules. Con- versely, in a corpus-based/statistical MT approach This paper presents a shallow-transfer ma- (Koehn, 2010), no such effort is required as the sys- chine translation (MT) system for translat- tem can be automatically built from parallel corpora ing from Kazakh to Turkish.
    [Show full text]
  • Apertium: a Rule-Based Machine Translation Platform
    Apertium: A rule-based machine translation platform Francis M. Tyers HSL-fakultehta UiT Norgga árktalaš universitehta 9018 Romsa (Norway) 28th September 2014 1 / 41 Outline Introduction Design Development Status Teaching Courses Google Summer of Code Google Code-in Research New language pairs Applying unsupervised methods Hybrid systems Future work and challenges Challenges Plans Collaboration 1 / 41 History I 2004 — Spain gets a new government which launches a call for proposals to build machine translation systems for the languages of Spain I The Universitat d’Alacant (UA), in consortium with EHU, UPC, UVigo, Eleka, Elhuyar (Basque Country) and Imaxin (Galicia) get funded to develop two MT systems: I Apertium: Spanish–Catalan, Spanish–Galician I Matxin: Spanish!Basque I Apertium was not built from scratch, but was rather a rewrite of two existing closed-source systems which had been built by UA 2 / 41 Focus I Marginalised: Languages which are on the periphery either societally or technologically (from Breton to Bengali). I Lesser-resourced: Languages for which few free/open-source language resources exist. I Closely-related: Languages which are suited to shallow-transfer machine translation. 3 / 41 Translation philosophy I Build on word-for-word translation I Avoid complex linguistic formalisms I We like saying “only secondary-school linguistics required”1 I Intermediate representation based on morphological information I Transformer-style systems I Analyse the source language (SL), then I Apply rules to ‘transform’ the SL to the TL 1Otte, P. and Tyers, F. (2011) “Rapid rule-based machine translation between Dutch and Afrikaans”. Proceedings of the 16th Annual Conference of the European Association of Machine Translation 4 / 41 Pipeline SL deformatter text morph.
    [Show full text]
  • The Apertium Bilingual Dictionaries on the Web of Data
    Semantic Web 0 (0) 1 1 IOS Press The Apertium Bilingual Dictionaries on the Web of Data Editor(s): Philipp Cimiano, Universität Bielefeld, Germany Solicited review(s): Roberto Navigli, Sapienza University of Rome, Italy; Bettina Klimek, University of Leipzig, Germany; Sebastian Hellmann, University of Leipzig, Gemany; one anonymous reviewer Jorge Gracia a;∗, Marta Villegas b, Asunción Gómez-Pérez a, and Núria Bel b, a Ontology Engineering Group, Universidad Politécnica de Madrid Campus de Montegancedo s/n Boadilla del Monte 28660 Madrid. Spain E-mail: {jgracia,asun}@fi.upm.es b Institut Universitari Linguistica Aplicada, Universitat Pompeu Fabra Roc Boronat, 138 08018 Barcelona. Spain E-mail: {marta.villegas,nuria.bel}@upf.edu Abstract. Bilingual electronic dictionaries contain collections of lexical entries in two languages, with explicitly declared trans- lation relations between such entries. Nevertheless, they are typically developed in isolation, in their own formats and accessible through proprietary APIs. In this paper we propose the use of Semantic Web techniques to make translations available on the Web to be consumed by other semantic enabled resources in a direct manner, based on standard languages and query means. In particular, we describe the conversion of the Apertium family of bilingual dictionaries and lexicons into RDF (Resource De- scription Framework) and how their data have been made accessible on the Web as linked data. As a result, all the converted dictionaries (many of them covering under-resourced languages) are connected among them and can be easily traversed from one to another to obtain, for instance, translations between language pairs not originally connected in any of the original dictionaries.
    [Show full text]
  • New Inflectional Lexicons and Training Corpora for Improved
    New Inflectional Lexicons and Training Corpora for Improved Morphosyntactic Annotation of Croatian and Serbian Nikola Ljubesiˇ c´∗ Filip Klubickaˇ ∗ Zeljkoˇ Agic´y Ivo-Pavao Jazbecz * Dept. of Information and Communication Sciences, University of Zagreb Ivana Luciˇ ca´ 3, HR-10000 Zagreb, Croatia y Center for Language Technology, University of Copenhagen Njalsgade 140, 2300 Copenhagen S, Denmark z Institute of Croatian Language and Linguistics Ulica Republike Austrije 16, HR-10000 Zagreb, Croatia fnljubesi,[email protected] [email protected] [email protected] Abstract In this paper we present newly developed inflectional lexcions and manually annotated corpora of Croatian and Serbian. We introduce hrLex and srLex—two freely available inflectional lexicons of Croatian and Serbian—and describe the process of building these lexicons, supported by supervised machine learning techniques for lemma and paradigm prediction. Furthermore, we introduce hr500k, a manually annotated corpus of Croatian, 500 thousand tokens in size. We showcase the three newly developed resources on the task of morphosyntactic annotation of both languages by using a recently developed CRF tagger. We achieve best results yet reported on the task for both languages, beating the HunPos baseline trained on the same datasets by a wide margin. Keywords: inflectional lexicon, morphosyntactic annotation, Croatian, Serbian 1. Introduction expansion or enrichment. Snajderˇ et al. (2008) provide an- other line of work on Croatian inflectional lexica, but the In this paper we introduce hrLex and srLex, freely available resulting resource is not freely available. morphological lexicons of Croatian and Serbian. We de- For Serbian the SrpMD dictionary (Krstev, 2008), 85,721 scribe the process of building these lexicons which includes lemmas in size, is published under a non-commercial li- supervised machine learning techniques for lemma and cense and indexed on Meta-Share, but is not available for paradigm candidate ranking of non-covered words to en- download.
    [Show full text]