Cross-Lingual Terminology Extraction for Translation Quality Estimation⋆
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
A Comparison of Knowledge Extraction Tools for the Semantic Web
A Comparison of Knowledge Extraction Tools for the Semantic Web Aldo Gangemi1;2 1 LIPN, Universit´eParis13-CNRS-SorbonneCit´e,France 2 STLab, ISTC-CNR, Rome, Italy. Abstract. In the last years, basic NLP tasks: NER, WSD, relation ex- traction, etc. have been configured for Semantic Web tasks including on- tology learning, linked data population, entity resolution, NL querying to linked data, etc. Some assessment of the state of art of existing Knowl- edge Extraction (KE) tools when applied to the Semantic Web is then desirable. In this paper we describe a landscape analysis of several tools, either conceived specifically for KE on the Semantic Web, or adaptable to it, or even acting as aggregators of extracted data from other tools. Our aim is to assess the currently available capabilities against a rich palette of ontology design constructs, focusing specifically on the actual semantic reusability of KE output. 1 Introduction We present a landscape analysis of the current tools for Knowledge Extraction from text (KE), when applied on the Semantic Web (SW). Knowledge Extraction from text has become a key semantic technology, and has become key to the Semantic Web as well (see. e.g. [31]). Indeed, interest in ontology learning is not new (see e.g. [23], which dates back to 2001, and [10]), and an advanced tool like Text2Onto [11] was set up already in 2005. However, interest in KE was initially limited in the SW community, which preferred to concentrate on manual design of ontologies as a seal of quality. Things started changing after the linked data bootstrapping provided by DB- pedia [22], and the consequent need for substantial population of knowledge bases, schema induction from data, natural language access to structured data, and in general all applications that make joint exploitation of structured and unstructured content. -
A Combined Method for E-Learning Ontology Population Based on NLP and User Activity Analysis
A Combined Method for E-Learning Ontology Population based on NLP and User Activity Analysis Dmitry Mouromtsev, Fedor Kozlov, Liubov Kovriguina and Olga Parkhimovich ITMO University, St. Petersburg, Russia [email protected], [email protected], [email protected], [email protected] Abstract. The paper describes a combined approach to maintaining an E-Learning ontology in dynamic and changing educational environment. The developed NLP algorithm based on morpho-syntactic patterns is ap- plied for terminology extraction from course tasks that allows to interlink extracted terms with the instances of the system’s ontology whenever some educational materials are changed. These links are used to gather statistics, evaluate quality of lectures’ and tasks’ materials, analyse stu- dents’ answers to the tasks and detect difficult terminology of the course in general (for the teachers) and its understandability in particular (for every student). Keywords: Semantic Web, Linked Learning, terminology extraction, education, educational ontology population 1 Introduction Nowadays reusing online educational resources becomes one of the most promis- ing approaches for e-learning systems development. A good example of using semantics to make education materials reusable and flexible is the SlideWiki system[1]. The key feature of an ontology-based e-learning system is the possi- bility for tutors and students to treat elements of educational content as named objects and named relations between them. These names are understandable both for humans as titles and for the system as types of data. Thus educational materials in the e-learning system thoroughly reflect the structure of education process via relations between courses, modules, lectures, tests and terms. -
Information Extraction Using Natural Language Processing
INFORMATION EXTRACTION USING NATURAL LANGUAGE PROCESSING Cvetana Krstev University of Belgrade, Faculty of Philology Information Retrieval and/vs. Natural Language Processing So close yet so far Outline of the talk ◦ Views on Information Retrieval (IR) and Natural Language Processing (NLP) ◦ IR and NLP in Serbia ◦ Language Resources (LT) in the core of NLP ◦ at University of Belgrade (4 representative resources) ◦ LR and NLP for Information Retrieval and Information Extraction (IE) ◦ at University of Belgrade (4 representative applications) Wikipedia ◦ Information retrieval ◦ Information retrieval (IR) is the activity of obtaining information resources relevant to an information need from a collection of information resources. Searches can be based on full-text or other content-based indexing. ◦ Natural Language Processing ◦ Natural language processing is a field of computer science, artificial intelligence, and computational linguistics concerned with the interactions between computers and human (natural) languages. As such, NLP is related to the area of human–computer interaction. Many challenges in NLP involve: natural language understanding, enabling computers to derive meaning from human or natural language input; and others involve natural language generation. Experts ◦ Information Retrieval ◦ As an academic field of study, Information Retrieval (IR) is finding material (usually documents) of an unstructured nature (usually text) that satisfies an information need from within large collection (usually stored on computers). ◦ C. D. Manning, P. Raghavan, H. Schutze, “Introduction to Information Retrieval”, Cambridge University Press, 2008 ◦ Natural Language Processing ◦ The term ‘Natural Language Processing’ (NLP) is normally used to describe the function of software or hardware components in computer system which analyze or synthesize spoken or written language. -
A Terminology Extraction System for Technology Related Terms
TEST: A Terminology Extraction System for Technology Related Terms Murhaf Hossari Soumyabrata Dev John D. Kelleher ADAPT Centre, Trinity College ADAPT Centre, Trinity College ADAPT Centre, Technological Dublin Dublin University Dublin Dublin, Ireland Dublin, Ireland Dublin, Ireland [email protected] [email protected] [email protected] ABSTRACT of applications [6], ranging from e-mail spam filtering, advertising Tracking developments in the highly dynamic data-technology and marketing industry [4, 7], and social media. Particularly, in landscape are vital to keeping up with novel technologies and tools, these realm of online articles and blog articles, the growth in the in the various areas of Artificial Intelligence (AI). However, It is amount of information is almost in an exponential nature. Therefore, difficult to keep track of all the relevant technology keywords. In it is important for the researchers to develop a system that can this paper, we propose a novel system that addresses this problem. automatically parse the millions of documents, and identify the key This tool is used to automatically detect the existence of new tech- technological terms. Such system will greatly reduce the manual nologies and tools in text, and extract terms used to describe these work, and save several man-hours. new technologies. The extracted new terms can be logged as new In this paper, we develop a machine-learning based solution, AI technologies as they are found on-the-fly in the web. It can be that is capable of the discovery and extraction of information re- subsequently classified into the relevant semantic labels and AIdo- lated to AI technologies. -
Ontology Learning and Its Application to Automated Terminology Translation
Natural Language Processing Ontology Learning and Its Application to Automated Terminology Translation Roberto Navigli and Paola Velardi, Università di Roma La Sapienza Aldo Gangemi, Institute of Cognitive Sciences and Technology lthough the IT community widely acknowledges the usefulness of domain ontolo- A gies, especially in relation to the Semantic Web,1,2 we must overcome several bar- The OntoLearn system riers before they become practical and useful tools. Thus far, only a few specific research for automated ontology environments have ontologies. (The “What Is an Ontology?” sidebar on page 24 provides learning extracts a definition and some background.) Many in the and is part of a more general ontology engineering computational-linguistics research community use architecture.4,5 Here, we describe the system and an relevant domain terms WordNet,3 but large-scale IT applications based on experiment in which we used a machine-learned it require heavy customization. tourism ontology to automatically translate multi- from a corpus of text, Thus, a critical issue is ontology construction— word terms from English to Italian. The method can identifying, defining, and entering concept defini- apply to other domains without manual adaptation. relates them to tions. In large, complex application domains, this task can be lengthy, costly, and controversial, because peo- OntoLearn architecture appropriate concepts in ple can have different points of view about the same Figure 1 shows the elements of the architecture. concept. Two main approaches aid large-scale ontol- Using the Ariosto language processor,6 OntoLearn a general-purpose ogy construction. The first one facilitates manual extracts terminology from a corpus of domain text, ontology engineering by providing natural language such as specialized Web sites and warehouses or doc- ontology, and detects processing tools, including editors, consistency uments exchanged among members of a virtual com- checkers, mediators to support shared decisions, and munity. -
The Linguistic Dimension of Terminology
1st Athens International Conference on Translation and Interpretation Translation: Between Art and Social Science, 13 -14 October 2006 THE LINGUISTIC DIMENSION OF TERMINOLOGY: PRINCIPLES AND METHODS OF TERM FORMATION Kostas Valeontis Elena Mantzari Physicist-Electronic Engineer, President of ΕLΕΤΟ1, Linguist-Researcher, Deputy Secretary General of ΕLΕΤΟ Abstract Terminology has a twofold meaning: 1. it is the discipline concerned with the principles and methods governing the study of concepts and their designations (terms, names, symbols) in any subject field, and the job of collecting, processing, and managing relevant data, and 2. the set of terms belonging to the special language of an individual subject field. In its study of concepts and their representations in special languages, terminology is multidisciplinary, since it borrows its fundamental tools and concepts from a number of disciplines (e.g. logic, ontology, linguistics, information science and other specific fields) and adapts them appropriately in order to cover particularities in its own area. The interdisciplinarity of terminology results from the multifaceted character of terminological units, as linguistic items (linguistics), as conceptual elements (logic, ontology, cognitive sciences) and as vehicles of communication in both scientific and generic language contexts. Accordingly, the theory of terminology can be identified as having three different dimensions: the cognitive, the linguistic, and the communicative dimension (Sager: 1990). The linguistic dimension of the theory of terminology can be detected mainly in the linguistic mechanisms that set the patterns for term formation and term forms. In this paper, we will focus on the presentation of general linguistic principles concerning term formation, during primary naming of an original concept in a source language and secondary term formation in a target language. -
Translate's Localization Guide
Translate’s Localization Guide Release 0.9.0 Translate Jun 26, 2020 Contents 1 Localisation Guide 1 2 Glossary 191 3 Language Information 195 i ii CHAPTER 1 Localisation Guide The general aim of this document is not to replace other well written works but to draw them together. So for instance the section on projects contains information that should help you get started and point you to the documents that are often hard to find. The section of translation should provide a general enough overview of common mistakes and pitfalls. We have found the localisation community very fragmented and hope that through this document we can bring people together and unify information that is out there but in many many different places. The one section that we feel is unique is the guide to developers – they make assumptions about localisation without fully understanding the implications, we complain but honestly there is not one place that can help give a developer and overview of what is needed from them, we hope that the developer section goes a long way to solving that issue. 1.1 Purpose The purpose of this document is to provide one reference for localisers. You will find lots of information on localising and packaging on the web but not a single resource that can guide you. Most of the information is also domain specific ie it addresses KDE, Mozilla, etc. We hope that this is more general. This document also goes beyond the technical aspects of localisation which seems to be the domain of other lo- calisation documents. -
File: Terminology.Pdf
Lowe I 2009, www.scientificlanguage.com/esp/terminology.pdf 1 A question of terminology ______________________________________________________________ This essay is copyright under the creative commons Attribution - Noncommercial - Share Alike 3.0 Unported licence. You are encouraged to share and copy this essay free of charge. See for details: http://creativecommons.org/licenses/by-nc-sa/3.0/ A question of terminology Updated 24 November 2009 24 November 2009 A framework for analysing sub/semi-technical words is presented which reconciles the two senses of these terms. Additional practical ideas for teaching them are provided. 0. Introduction The terminology for describing the language of science is in a state of confusion. There are several words, with differing definitions. There is even confusion over exactly what should be defined and labelled. In other words, there is a problem of classification. This paper attempts to sort out some of the confusion. 1. Differing understandings of what is meant by “general” English and “specialised” English (see also www.scientificlanguage.com/esp/audience.pdf ) Table to show the different senses of ‘general’ and ‘specialised’ Popular senses Technical senses 1. a wide range of general 2. Basic English, of all interest topics of general kinds, from pronunciation, “general” English knowledge, such as sport, through vocabulary, to hobbies etc discourse patterns such as Suggested term: those used in a newspaper. general topics English Suggested terms: foundational English basic English 3. a wide range of topics 4. Advanced English, the within the speciality of the fine points, and the “specialised” English student. vocabulary and discourse Suggested term: patterns specific to the specialised topics English discipline. -
Generating Domain Terminologies Using Root- and Rule-Based Terms1
Generating Domain Terminologies using Root- and Rule-Based Terms1 Jacob Collard1, T. N. Bhat2, Eswaran Subrahmanian3,4, Ram D. Sriram3, John T. Elliot2, Ursula R. Kattner2, Carelyn E. Campbell2, Ira Monarch4,5 Independent Consultant, Ithaca, New York1, Materials Measurement Laboratory, National Institute of Standards and Technology, Gaithersburg, MD2, Information Technology Laboratory, National Institute of Standards and Technology, Gaithersburg, MD 3, Carnegie Mellon University, Pittsburgh, PA4, Independent Consultant, Pittsburgh, PA5 Abstract Motivated by the need for flexible, intuitive, reusable, and normalized terminology for guiding search and building ontologies, we present a general approach for generating sets of such terminologies from natural language documents. The terms that this approach generates are root- and rule-based terms, generated by a series of rules designed to be flexible, to evolve, and, perhaps most important, to protect against ambiguity and standardize semantically similar but syntactically distinct phrases to a normal form. This approach combines several linguistic and computational methods that can be automated with the help of training sets to quickly and consistently extract normalized terms. We discuss how this can be extended as natural language technologies improve and how the strategy applies to common use-cases such as search, document entry and archiving, and identifying, tracking, and predicting scientific and technological trends. Keywords dependency parsing; natural language processing; ontology generation; search; terminology generation; unsupervised learning. 1. Introduction 1.1 Terminologies and Semantic Technologies Services and applications on the world-wide web, as well as standards defined by the World Wide Web Consortium (W3C), the primary standards organization for the web, have been integrating semantic technologies into the Internet since 2001 (Koivunen and Miller 2001). -
Facilitating Terminology Translation with Target Lemma Annotations
Facilitating Terminology Translation with Target Lemma Annotations Toms Bergmanisyz and Marcis¯ Pinnisyz yTilde / Vien¯ıbas gatve 75A, Riga, Latvia zFaculty of Computing, University of Latvia / Rain¸a bulv. 19, Riga, Latvia [email protected] Abstract has assumed that the correct morphological forms are apriori known (Hokamp and Liu, 2017; Post Most of the recent work on terminology in- and Vilar, 2018; Hasler et al., 2018; Dinu et al., tegration in machine translation has assumed 2019; Song et al., 2020; Susanto et al., 2020; Dou- that terminology translations are given already gal and Lonsdale, 2020). Thus previous work has inflected in forms that are suitable for the tar- get language sentence. In day-to-day work of approached terminology translation predominantly professional translators, however, it is seldom as a problem of making sure that the decoder’s the case as translators work with bilingual glos- output contains lexically and morphologically pre- saries where terms are given in their dictionary specified target language terms. While useful in forms; finding the right target language form some cases and some languages, such approaches is part of the translation process. We argue come short of addressing terminology translation that the requirement for apriori specified tar- into morphologically complex languages where get language forms is unrealistic and impedes the practical applicability of previous work. In each word can have many morphological surface this work, we propose to train machine trans- forms. lation systems using a source-side data aug- For terminology translation to be viable for trans- mentation method1 that annotates randomly se- lation into morphologically complex languages, lected source language words with their tar- terminology constraints have to be soft. -
Terminology Extraction, Translation Tools and Comparable Corpora
Terminology Extraction, Translation Tools and Comparable Corpora: TTC concept, midterm progress and achieved results Tatiana Gornostay, Anita Gojun, Marion Weller, Ulrich Heid, Emmanuel Morin, Beatrice Daille, Helena Blancafort, Serge Sharoff, Claude Méchoulam To cite this version: Tatiana Gornostay, Anita Gojun, Marion Weller, Ulrich Heid, Emmanuel Morin, et al.. Terminology Extraction, Translation Tools and Comparable Corpora: TTC concept, midterm progress and achieved results. LREC 2012 Workshop on Creating Cross-language Resources for Disconnected Languages and Styles (CREDISLAS), May 2012, Istanbul, Turkey. 4 p. hal-00819909 HAL Id: hal-00819909 https://hal.archives-ouvertes.fr/hal-00819909 Submitted on 9 May 2013 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Terminology Extraction, Translation Tools and Comparable Corpora: TTC concept, midterm progress and achieved results Tatiana Gornostaya, Anita Gojunb, Marion Wellerb, Ulrich Heidb, Emmanuel Morinc, Beatrice Daillec, Helena Blancafortd, Serge Sharoffe, Claude Méchoulamf Tildea, Institute -
Terminology, Phraseology, and Lexicography 1. Introduction Sinclair
Terminology, Phraseology, and Lexicography1 Patrick Hanks Institute of Formal and Applied Linguistics, Charles University in Prague This paper explores two aspects of word use and word meaning in terms of Sinclair's (1991, 1998) distinction between the open-choice principle (or terminological tendency) and the idiom principle (or phraseological tendency). Technical terms such as strobilation are rare, highly domain-specific, and of little phraseological interest, although the texts in which such word occur do tend to contain interesting clusters of domain-specific terminology. At the other extreme, it is impossible to know the meaning of ordinary common words such as the verb blow without knowing the phraseological context in which the word is used. Many words have both a terminological tendency and a phraseological tendency. In some cases the two tendencies are in harmony; in other cases there is tension between them. The relationship between these two tendencies is investigated, using examples from the British National Corpus. 1. Introduction Sinclair (1991) makes a distinction between two aspects of meaning in texts, which he calls the open-choice principle and the idiom principle. In this paper, I will try to show why it is of great importance, not only for understanding meaning in text, but also for lexicography, to try to take account of both principles even-handedly and assess the balance between them when analyzing the meaning and use of a word. The open-choice principle (alternatively called the terminological tendency) is: a way of seeing language as the result of a very large number of complex choices. At each point where a unit is complete (a word or a phrase or a clause), a large range of choices opens up and the only restraint is grammaticalness.