EG 202 421 V1.1.1 (2006-11) ETSI Guide

Total Page:16

File Type:pdf, Size:1020Kb

EG 202 421 V1.1.1 (2006-11) ETSI Guide Final draft ETSI EG 202 421 V1.1.1 (2006-11) ETSI Guide Human Factors (HF); Multicultural and language aspects of multimedia communications 2 Final draft ETSI EG 202 421 V1.1.1 (2006-11) Reference DEG/HF-00054 Keywords broadband, ID, multimedia, user ETSI 650 Route des Lucioles F-06921 Sophia Antipolis Cedex - FRANCE Tel.: +33 4 92 94 42 00 Fax: +33 4 93 65 47 16 Siret N° 348 623 562 00017 - NAF 742 C Association à but non lucratif enregistrée à la Sous-Préfecture de Grasse (06) N° 7803/88 Important notice Individual copies of the present document can be downloaded from: http://www.etsi.org The present document may be made available in more than one electronic version or in print. In any case of existing or perceived difference in contents between such versions, the reference version is the Portable Document Format (PDF). In case of dispute, the reference shall be the printing on ETSI printers of the PDF version kept on a specific network drive within ETSI Secretariat. Users of the present document should be aware that the document may be subject to revision or change of status. Information on the current status of this and other ETSI documents is available at http://portal.etsi.org/tb/status/status.asp If you find errors in the present document, please send your comment to one of the following services: http://portal.etsi.org/chaircor/ETSI_support.asp Copyright Notification No part may be reproduced except as authorized by written permission. The copyright and the foregoing restriction extend to reproduction in all media. © European Telecommunications Standards Institute 2006. All rights reserved. DECTTM, PLUGTESTSTM and UMTSTM are Trade Marks of ETSI registered for the benefit of its Members. TIPHONTM and the TIPHON logo are Trade Marks currently being registered by ETSI for the benefit of its Members. 3GPPTM is a Trade Mark of ETSI registered for the benefit of its Members and of the 3GPP Organizational Partners. ETSI 3 Final draft ETSI EG 202 421 V1.1.1 (2006-11) Contents Intellectual Property Rights................................................................................................................................6 Foreword.............................................................................................................................................................6 Introduction ........................................................................................................................................................6 1 Scope ........................................................................................................................................................8 2 References ................................................................................................................................................9 3 Definitions and abbreviations.................................................................................................................12 3.1 Definitions........................................................................................................................................................12 3.2 Abbreviations ...................................................................................................................................................13 4 Language and cultural variation.............................................................................................................14 4.1 Diversity of languages and cultures..................................................................................................................14 4.2 Internationalization and localization ................................................................................................................14 4.2.1 Translation memory....................................................................................................................................15 4.3 Current ways of providing cultural and language variants...............................................................................15 4.4 Text to speech...................................................................................................................................................16 4.4.1 Text to speech (TTS) engines .....................................................................................................................16 4.4.2 Speech Synthesis Mark-up Language (SSML)...........................................................................................17 4.4.3 Spoken Text Mark-up Language (STML) ..................................................................................................18 4.4.4 Java Speech Mark-up Language (JSML)....................................................................................................18 4.4.5 SABLE........................................................................................................................................................18 4.4.6 Voice Extensible Mark-up Language (VoiceXML) ...................................................................................18 4.5 Speech recognition ...........................................................................................................................................18 4.6 Language identification....................................................................................................................................19 4.7 Design for flexible handling of users' needs.....................................................................................................20 4.7.1 Universal Communications Identifier (UCI) ..............................................................................................20 4.7.2 User profiles................................................................................................................................................20 4.8 European multicultural and multilingual research............................................................................................21 4.9 Existing standards, recommendations and guidelines ......................................................................................21 4.9.1 Localization standards ................................................................................................................................21 4.9.1.1 Terminology, lexicon and ontology ......................................................................................................22 4.9.1.2 Translation memory ..............................................................................................................................22 4.9.1.3 Localization markup language ..............................................................................................................23 4.9.1.4 Structured information ..........................................................................................................................23 4.9.2 Natural language representation .................................................................................................................23 4.9.3 Sign languages............................................................................................................................................24 4.9.4 Description of language and cultural capabilities .......................................................................................24 4.9.5 Character sets and character encoding........................................................................................................25 4.9.6 Sorting order...............................................................................................................................................25 4.9.7 Keyboard layouts ........................................................................................................................................25 4.9.8 Spoken command vocabulary.....................................................................................................................26 4.9.9 Mobile telephony product and service design.............................................................................................26 4.9.10 Smart cards.................................................................................................................................................27 4.9.11 User profiles................................................................................................................................................27 4.9.12 Universal Communications Identifier (UCI) ..............................................................................................28 4.10 Current constraints ...........................................................................................................................................28 4.10.1 Commercial.................................................................................................................................................29 4.10.2 Legal ...........................................................................................................................................................29 4.10.3 Technological and operational....................................................................................................................29 5 Requirements..........................................................................................................................................31 5.1 Introduction ......................................................................................................................................................31
Recommended publications
  • Can You Give Me Another Word for Hyperbaric?: Improving Speech
    “CAN YOU GIVE ME ANOTHER WORD FOR HYPERBARIC?”: IMPROVING SPEECH TRANSLATION USING TARGETED CLARIFICATION QUESTIONS Necip Fazil Ayan1, Arindam Mandal1, Michael Frandsen1, Jing Zheng1, Peter Blasco1, Andreas Kathol1 Fred´ eric´ Bechet´ 2, Benoit Favre2, Alex Marin3, Tom Kwiatkowski3, Mari Ostendorf3 Luke Zettlemoyer3, Philipp Salletmayr5∗, Julia Hirschberg4, Svetlana Stoyanchev4 1 SRI International, Menlo Park, USA 2 Aix-Marseille Universite,´ Marseille, France 3 University of Washington, Seattle, USA 4 Columbia University, New York, USA 5 Graz Institute of Technology, Austria ABSTRACT Our previous work on speech-to-speech translation systems has We present a novel approach for improving communication success shown that there are seven primary sources of errors in translation: between users of speech-to-speech translation systems by automat- • ASR named entity OOVs: Hi, my name is Colonel Zigman. ically detecting errors in the output of automatic speech recogni- • ASR non-named entity OOVs: I want some pristine plates. tion (ASR) and statistical machine translation (SMT) systems. Our approach initiates system-driven targeted clarification about errorful • Mispronunciations: I want to collect some +de-MOG-raf-ees regions in user input and repairs them given user responses. Our about your family? (demographics) system has been evaluated by unbiased subjects in live mode, and • Homophones: Do you have any patients to try this medica- results show improved success of communication between users of tion? (patients vs. patience) the system. • MT OOVs: Where is your father-in-law? Index Terms— Speech translation, error detection, error correc- • Word sense ambiguity: How many men are in your com- tion, spoken dialog systems. pany? (organization vs.
    [Show full text]
  • A Survey of Voice Translation Methodologies - Acoustic Dialect Decoder
    International Conference On Information Communication & Embedded Systems (ICICES-2016) A SURVEY OF VOICE TRANSLATION METHODOLOGIES - ACOUSTIC DIALECT DECODER Hans Krupakar Keerthika Rajvel Dept. of Computer Science and Engineering Dept. of Computer Science and Engineering SSN College Of Engineering SSN College of Engineering E-mail: [email protected] E-mail: [email protected] Bharathi B Angel Deborah S Vallidevi Krishnamurthy Dept. of Computer Science Dept. of Computer Science Dept. of Computer Science and Engineering and Engineering and Engineering SSN College Of Engineering SSN College Of Engineering SSN College Of Engineering E-mail: [email protected] E-mail: [email protected] E-mail: [email protected] Abstract— Speech Translation has always been about giving in order to make the process of communication amongst source text/audio input and waiting for system to give humans better, easier and efficient. However, the existing translated output in desired form. In this paper, we present the methods, including Google voice translators, typically Acoustic Dialect Decoder (ADD) – a voice to voice ear-piece handle the process of translation in a non-automated manner. translation device. We introduce and survey the recent advances made in the field of Speech Engineering, to employ in This makes the process of translation of word(s) and/or the ADD, particularly focusing on the three major processing sentences from one language to another, slower and more steps of Recognition, Translation and Synthesis. We tackle the tedious. We wish to make that process automatic – have a problem of machine understanding of natural language by device do what a human translator does, inside our ears.
    [Show full text]
  • Learning Speech Translation from Interpretation
    Learning Speech Translation from Interpretation zur Erlangung des akademischen Grades eines Doktors der Ingenieurwissenschaften von der Fakult¨atf¨urInformatik Karlsruher Institut f¨urTechnologie (KIT) genehmigte Dissertation von Matthias Paulik aus Karlsruhe Tag der m¨undlichen Pr¨ufung: 21. Mai 2010 Erster Gutachter: Prof. Dr. Alexander Waibel Zweiter Gutachter: Prof. Dr. Tanja Schultz Ich erkl¨arehiermit, dass ich die vorliegende Arbeit selbst¨andig verfasst und keine anderen als die angegebenen Quellen und Hilfsmittel verwendet habe sowie dass ich die w¨ortlich oder inhaltlich ¨ubernommenen Stellen als solche kenntlich gemacht habe und die Satzung des KIT, ehem. Universit¨atKarlsruhe (TH), zur Sicherung guter wissenschaftlicher Praxis in der jeweils g¨ultigenFassung beachtet habe. Karlsruhe, den 21. Mai 2010 Matthias Paulik Abstract The basic objective of this thesis is to examine the extent to which automatic speech translation can benefit from an often available but ignored resource, namely human interpreter speech. The main con- tribution of this thesis is a novel approach to speech translation development, which makes use of that resource. The performance of the statistical models employed in modern speech translation systems depends heavily on the availability of vast amounts of training data. State-of-the-art systems are typically trained on: (1) hundreds, sometimes thousands of hours of manually transcribed speech audio; (2) bi-lingual, sentence-aligned text corpora of man- ual translations, often comprising tens of millions of words; and (3) monolingual text corpora, often comprising hundreds of millions of words. The acquisition of such enormous data resources is highly time-consuming and expensive, rendering the development of deploy- able speech translation systems prohibitive to all but a handful of eco- nomically or politically viable languages.
    [Show full text]
  • Access to Blissymbolics In
    Access to Blissymbolics in ICT state-of-the-art and visions Mats Lundälv, DART, Sahlgrenska University Hospital, and BCI, Göteborg, Sweden Presented by Stephen van Tetzchner, ISAAC Research Symposium, Pittsburgh, USA, 2012 Introduction: Blissymbolics is, unlike most other graphical AAC systems, well suited for use on all technology levels, from no-tech, over low-tech, to current hi-tech ICT platforms. Blisymbols may be drawn by hand, in the sand, with paper and pen, or on a blackboard or whiteboard. In current practice we are however, as we are for all other symbol resources, heavily dependent on ICT support for practically managing and using Blissymbolics. It may be for setting up and printing symbol charts for low-tech use, or for writing documents or communicating remotely via email etc on computers and mobile devices. This presentation is an attempt to give an overview of the availability of Blissymbolics on current ICT platforms 2012, and some hints about what we can expect for the near future. Background: Blissymbolics was probably the first graphical AAC system to be supported on the emerging computer platforms in the early 1980:s; Talking BlissApple was one of the first pieces of AAC software developed for Apple II computers by Gregg Vanderheiden et al at Trace R&D Center, Madison Wisconsin USA. It was a break-through and was followed by a large number of Blissymbol AAC programs for the different computing platforms over the next decades. However, when the real expansion of the AAC technology came, with a number of commercially promoted pictorial systems such as PCS etc, Blissymbolics was largely dropping out of support on the dominating AAC software and hardware platforms.
    [Show full text]
  • Is 42 the Answer to Everything in Subtitling-Oriented Speech Translation?
    Is 42 the Answer to Everything in Subtitling-oriented Speech Translation? Alina Karakanta Matteo Negri Marco Turchi Fondazione Bruno Kessler Fondazione Bruno Kessler Fondazione Bruno Kessler University of Trento Trento - Italy Trento - Italy Trento - Italy [email protected] [email protected] [email protected] Abstract content, still rely heavily on human effort. In a typi- cal multilingual subtitling workflow, a subtitler first Subtitling is becoming increasingly important creates a subtitle template (Georgakopoulou, 2019) for disseminating information, given the enor- by transcribing the source language audio, timing mous amounts of audiovisual content becom- and adapting the text to create proper subtitles in ing available daily. Although Neural Machine Translation (NMT) can speed up the process the source language. These source language subti- of translating audiovisual content, large man- tles (also called captions) are already compressed ual effort is still required for transcribing the and segmented to respect the subtitling constraints source language, and for spotting and seg- of length, reading speed and proper segmentation menting the text into proper subtitles. Cre- (Cintas and Remael, 2007; Karakanta et al., 2019). ating proper subtitles in terms of timing and In this way, the work of an NMT system is already segmentation highly depends on information simplified, since it only needs to translate match- present in the audio (utterance duration, natu- ing the length of the source text (Matusov et al., ral pauses). In this work, we explore two meth- ods for applying Speech Translation (ST) to 2019; Lakew et al., 2019). However, the essence subtitling: a) a direct end-to-end and b) a clas- of a good subtitle goes beyond matching a prede- sical cascade approach.
    [Show full text]
  • Speech Translation Into Pakistan Sign Language
    Master Thesis Computer Science Thesis No: MCS-2010-24 2012 Speech Translation into Pakistan Sign Language Speech Translation into Pakistan Sign AhmedLanguage Abdul Haseeb Asim Illyas Speech Translation into Pakistan Sign LanguageAsim Ilyas School of Computing Blekinge Institute of Technology SE – 371 79 Karlskrona Sweden This thesis is submitted to the School of Computing at Blekinge Institute of Technology in partial fulfillment of the requirements for the degree of Master of Science in Computer Science. The thesis is equivalent to 20 weeks of full time studies. Contact Information: Authors: Ahmed Abdul Haseeb E-mail: [email protected] Asim Ilyas E-mail: [email protected] University advisor: Prof. Sara Eriksén School of Computing, Blekinge Institute of Technology Internet : www.bth.se/com SE – 371 79 Karlskrona Phone : +46 457 38 50 00 Sweden Fax : + 46 457 271 25 ABSTRACT Context: Communication is a primary human need and language is the medium for this. Most people have the ability to listen and speak and they use different languages like Swedish, Urdu and English etc. to communicate. Hearing impaired people use signs to communicate. Pakistan Sign Language (PSL) is the preferred language of the deaf in Pakistan. Currently, human PSL interpreters are required to facilitate communication between the deaf and hearing; they are not always available, which means that communication among the deaf and other people may be impaired or nonexistent. In this situation, a system with voice recognition as an input and PSL as an output will be highly helpful. Objectives: As part of this thesis, we explore challenges faced by deaf people in everyday life while interacting with unimpaired.
    [Show full text]
  • Application of Blissymbolics to the Non-Vocal Communicatively Handicapped
    University of Montana ScholarWorks at University of Montana Graduate Student Theses, Dissertations, & Professional Papers Graduate School 1977 Application of Blissymbolics to the non-vocal communicatively handicapped Karen Lynn Jones The University of Montana Follow this and additional works at: https://scholarworks.umt.edu/etd Let us know how access to this document benefits ou.y Recommended Citation Jones, Karen Lynn, "Application of Blissymbolics to the non-vocal communicatively handicapped" (1977). Graduate Student Theses, Dissertations, & Professional Papers. 1576. https://scholarworks.umt.edu/etd/1576 This Thesis is brought to you for free and open access by the Graduate School at ScholarWorks at University of Montana. It has been accepted for inclusion in Graduate Student Theses, Dissertations, & Professional Papers by an authorized administrator of ScholarWorks at University of Montana. For more information, please contact [email protected]. THE APPLICATION OF BLISSYMBOLICS TO THE NON-VOCAL COMMUNICATIVELY HANDICAPPED by Karen L. Jones B.A., University of Montana, 1974 Presented in partial fulfillment of the requirements for the degree of Master of Communication Sciences and Disorders UNIVERSITY OF MONTANA 1977 Approved by: Chairman iBo^d of ^miners Deaw^ Gradua t^chool Date UMI Number; EP34649 All rights reserved INFORMATION TO ALL USERS The quality of this reproduction is dependent on the quality of the copy submitted. In the unlikely event that the author did not send a complete manuscript and there are missing pages, these will be noted. Also, if material had to be removed, a note will indicate the deletion. UMT IXMHtitian PUbMIng UMI EP34649 Copyright 2012 by ProQuest LLC. All rights reserved.
    [Show full text]
  • The Fundamental Rules of Blissymbolics: Creating New Blissymbolics Characters and Vocabulary
    The fundamental rules of Blissymbolics: creating new Blissymbolics characters and vocabulary Blissymbolics Communication International (BCI) ¯ 2009-01-26 1 Introduction 1 2 Blissymbolics 2 3 Definitions 2 4 Graphic aspects of the language 4 5 Bliss-characters 7 6 Bliss-words 11 7 Indicators 16 8 Wordbuilding strategies for vocabulary extension 19 9 The Blissymbolics Development Process 22 10 Bibliography 26 11 The history of Blissymbol standardization 26 12 Figure 1: Summary of the Blissymbolics development process 29 13 Figure 2: Flow chart of the Blissymbolics development process 30 1.0 Introduction. This document describes the basic structure of the Blissymbolics language, and outlines both the rules necessary to be followed for creating new BCI Authorized Vocabulary, as well as procedures used for adopting that vocabulary. This reference document will guide anyone wishing to use the Blissymbolics language. Its purpose is to ensure consistency and maintain the integrity of Blissymbolics as an international language. The formal process for the development of Blissymbolics is outlined in clause 9. NOTE: A number of technical notes appear throughout the document in smaller type. These notes refer to a number of elements which are technical in nature, such as providing specific advice for font implementations (clause 4.3.6) or the need to keep the creation of new Bliss-characters to a minimum (clause 8.9). Many users of this document will not need to take these notes into account for purposes of teaching, but they are nonetheless important for vocabulary development work and do form a part of the official guidelines. 1.1 Target users.
    [Show full text]
  • Between Flexibility and Consistency: Joint Generation of Captions And
    Between Flexibility and Consistency: Joint Generation of Captions and Subtitles Alina Karakanta1,2, Marco Gaido1,2, Matteo Negri1, Marco Turchi1 1 Fondazione Bruno Kessler, Via Sommarive 18, Povo, Trento - Italy 2 University of Trento, Italy {akarakanta,mgaido,negri,turchi}@fbk.eu Abstract material but also between each other, for example in the number of blocks (pieces of time-aligned Speech translation (ST) has lately received text) they occupy, their length and segmentation. growing interest for the generation of subtitles Consistency is vital for user experience, for ex- without the need for an intermediate source language transcription and timing (i.e. cap- ample in order to elicit the same reaction among tions). However, the joint generation of source multilingual audiences, or to facilitate the quality captions and target subtitles does not only assurance process in the localisation industry. bring potential output quality advantageswhen Previous work in ST for subtitling has focused the two decoding processes inform each other, on generating interlingual subtitles (Matusov et al., but it is also often required in multilingual sce- 2019; Karakanta et al., 2020a), a) without consid- narios. In this work, we focus on ST models ering the necessity of obtaining captions consis- which generate consistent captions-subtitles in tent with the target subtitles, and b) without ex- terms of structure and lexical content. We further introduce new metrics for evaluating amining whether the joint generation leads to im- subtitling consistency. Our findings show that provements in quality. We hypothesise that knowl- joint decoding leads to increased performance edge sharing between the tasks of transcription and consistency between the generated cap- and translation could lead to such improvements.
    [Show full text]
  • ONIX for Books Codelists Issue 40
    ONIX for Books Codelists Issue 40 23 January 2018 DOI: 10.4400/akjh All ONIX standards and documentation – including this document – are copyright materials, made available free of charge for general use. A full license agreement (DOI: 10.4400/nwgj) that governs their use is available on the EDItEUR website. All ONIX users should note that this is the fourth issue of the ONIX codelists that does not include support for codelists used only with ONIX version 2.1. Of course, ONIX 2.1 remains fully usable, using Issue 36 of the codelists or earlier. Issue 36 continues to be available via the archive section of the EDItEUR website (http://www.editeur.org/15/Archived-Previous-Releases). These codelists are also available within a multilingual online browser at https://ns.editeur.org/onix. Codelists are revised quarterly. Go to latest Issue Layout of codelists This document contains ONIX for Books codelists Issue 40, intended primarily for use with ONIX 3.0. The codelists are arranged in a single table for reference and printing. They may also be used as controlled vocabularies, independent of ONIX. This document does not differentiate explicitly between codelists for ONIX 3.0 and those that are used with earlier releases, but lists used only with earlier releases have been removed. For details of which code list to use with which data element in each version of ONIX, please consult the main Specification for the appropriate release. Occasionally, a handful of codes within a particular list are defined as either deprecated, or not valid for use in a particular version of ONIX or with a particular data element.
    [Show full text]
  • Of ISO/IEC 10646 and Unicode
    ISO/IEC JTC1/SC2/WG2 N2114 Title: Graphic representation of the Roadmap to the SMP, Plane 1 of the UCS Source: Ad hoc group on Roadmap Status: Expert contribution Date: 1999-09-15 Action: For confirmation by ISO/IEC JTC1/SC2/WG2 Replaces: N2046 The following tables comprise a real-size map of Plane 1, the SMP (Secondary Multilingual Plane) of the UCS (Universal Character Set). To print the HTML document it may be necessary to set the print percentage to 90% as the tables are wider than A4 or US Letter paper. The tables are formatted to use the Times font. The following conventions are used in the table to help the user identify the status of (colours can be seen in the online version of this document, http://www.dkuug.dk/jtc1/sc2/wg2/docs/n2114.pdf): Bold text indicates an allocated (i.e. published) character collection (none as yet in Plane 1). (Bold text between parentheses) indicates scripts which have been accepted for processing toward inclusion in the standard. (Text between parentheses) indicates scripts for which proposals have been submitted to WG2 or the UTC. ¿Text beween question marks? indicates scripts for which detailed proposals have not yet been written. ??? in a block indicates that no suggestion has been made regarding the block allocation. NOTE: With regard to the revision practice employed in this document, when scripts are actually proposed to WG2 or to the UTC, the practice is to "front" them in the zones to which they are tentatively allocated, and to adjust the block size with regard to the allocation proposed.
    [Show full text]
  • Implementing Machine Translation and Post-Editing to the Translation of Wildlife Documentaries Through Voice-Over and Off-Screen Dubbing
    ADVERTIMENT. Lʼaccés als continguts dʼaquesta tesi queda condicionat a lʼacceptació de les condicions dʼús establertes per la següent llicència Creative Commons: http://cat.creativecommons.org/?page_id=184 ADVERTENCIA. El acceso a los contenidos de esta tesis queda condicionado a la aceptación de las condiciones de uso establecidas por la siguiente licencia Creative Commons: http://es.creativecommons.org/blog/licencias/ WARNING. The access to the contents of this doctoral thesis it is limited to the acceptance of the use conditions set by the following Creative Commons license: https://creativecommons.org/licenses/?lang=en Universitat Autònoma de Barcelona Departament de Traducció i d’Interpretació i d’Estudis de l’Asia Oriental Doctorat en Traducció i Estudis Interculturals Implementing Machine Translation and Post-Editing to the Translation of Wildlife Documentaries through Voice-over and Off-screen Dubbing A Research on Effort and Quality PhD dissertation presented by: Carla Ortiz Boix Supervised and tutorized by: Dr. Anna Matamala 2016 A la meva família: als que hi són, als que no, i als que només hi són a mitges. Acknowledgments The road to finishing this PhD has not been easy and it would not have been accomplished without the priceless support of many: First of all, I want to thank my supervisor, Dr. Anna Matamala, for all her hard work. It has not been an easy road and sometimes I would have lost the right path if she had not been there to support, encourage, and challenge me. The PhD would not have come out the way it has without you. On a professional level, I also want to thank Dr.
    [Show full text]