Supervised Visualization of Vocabulary Knowledge Towards Explainable Support of Second Language Learners

Total Page:16

File Type:pdf, Size:1020Kb

Supervised Visualization of Vocabulary Knowledge Towards Explainable Support of Second Language Learners Supervised Visualization of Vocabulary Knowledge towards Explainable Support of Second Language Learners Yo Ehara Shizuoka Institute of Science and Technology, 2200-2, Toyosawa, Fukuroi, Shizuoka, Japan. [email protected] Abstract easy to understand and, hence, are well studied. However, in second language learning, most gaps are related to mean- In second language learning, it is crucial to identify gaps in ing and semantics and are inherently abstract. Hence, visu- knowledge of the language between second language learn- alizing these gaps is essential to make these gaps intuitively ers and native speakers. Such a gap exists even when learning a single word in a second language. As the semantic broad- understandable. ness of a word differs from language to language, language The broadness of a word, or how a word can be used in learners must learn how broadly a word can be used in a lan- the language to express different concepts, is one such ab- guage. For example, certain languages use different words for stract gap (Read 2000). Because the meaning of a word dif- “period” in “a period of time” or “period pains” yet both are fers from language to language, when learning a word in a nouns. Learners whose native languages are such languages second language, there typically exists a gap between what typically have only partial knowledge of a word, even though learners think the word means and how the word is actu- they think they know the word “period,” producing a gap be- ally used in the language. Polysemous words are examples tween them and native speakers. Language learners typically want explanations for these word usage differences, which that are easy to understand: “book” can mean an item asso- even native speakers find it difficult to explain and find it ciating with reading, or it can mean to make a reservation. costly to annotate. To support language learners in noticing Other than these examples, to which the part-of-speech tag- these challenging differences easily and intuitively, this pa- ging techniques in natural language processing (NLP) seem per proposes a novel supervised visualization of the usages applicable, some examples are more subtle: some languages of a word. In our method, the usages of an inputted word in always use different words for “time” in “in a short time” large corpora written by native speakers are visualized, tak- or “for a time,” in which the word “time” refers to a period, ing the semantic proximity between the usages into account. and “time and space” or “time heals all wounds,” in which Then, for the single inputted word, our method makes a per- “time” is used as an abstract concept. In another example, sonalized prediction of word usages that each learner may many languages use different words for “period” in “a pe- know, based on his/her results of a quick vocabulary test, which takes approximately 30 minutes. The experiment re- riod of time”, and “period” in “period pains”. In this way, sults show that our method produces better usage frequency the granularity of the word’s senses should be distinguished counts than raw usage frequency counts in predicting vocab- for second language acquisition, as it varies from word to ulary test responses, implying that word usage prediction is word. accurate. Polysemous words encode different concepts in one word: hence, they have been one of the central topics in knowl- Introduction edge engineering. A substantial amount of work has been conducted to automatically recognize polysemous words for Acquiring a second language requires repeated efforts to practical applications by using machine learning, includ- narrow the gap between language learners’ knowledge of the ing those in the previous AAAI-MAKE workshops (Ram- language and that of native speakers. Making such gaps intu- prasad and Maddox 2019; Hinkelmann et al. 2019; Lau- itively understandable greatly helps language learners self- renzi et al. 2019). However, even among few such applica- teach the language and also helps researchers build effec- tions for second language acquisition (Heilman et al. 2007; tive language tutoring systems. Some gaps such as vocabu- Dias and Moraliyski 2009) in the artificial intelligence (AI) lary size, or time spent in language learning are intuitively community, the challenging problem of different granularity of the word’s senses in second language acquisition has not Copyright c 2020 held by the author(s). In A. Martin, K. Hinkel- been addressed. In second language acquisition, as learners mann, H.-G. Fill, A. Gerber, D. Lenat, R. Stolle, F. van Harmelen (Eds.), Proceedings of the AAAI 2020 Spring Symposium on Com- are typically not linguistic experts, i.e., novices, hence, sys- bining Machine Learning and Knowledge Engineering in Practice tems to support their learning need to be intuitively under- (AAAI-MAKE 2020). Stanford University, Palo Alto, California, standable. Our goal is to make the gaps among word usages USA, March 23-25, 2020. Use permitted under Creative Commons intuitively understandable, even for novice language learn- License Attribution 4.0 International (CC BY 4.0). ers. To this end, this paper proposes a novel supervised visu- alization method for word usages to assist in learning the different usages of a word. Our method first searches all usages of the target word in a large corpus written by na- tive speakers. Then, it calculates the vector representation of each usage, or occurrence, of each word by using a contex- tualized word embedding method (Devlin et al. 2019). Con- textualized word embedding methods (Peters et al. 2018; Devlin et al. 2019) are recently proposed methods to embed each occurrence of a word, capturing the context of each us- age of the word. Then, our method is trained to visualize the contextual- ized word embedding vectors by plotting each usage as a point in a two-dimensional space. Unlike a typical visual- ization method that merely projects the vectors to a two- dimensional space, our method is trained to fit and visu- ally explain a given supervision dataset. This means that the same vectors are visualized in different ways if the su- pervision dataset differs. Here, the supervisions are a vo- cabulary test result dataset that consists of a matrix-format data, recording which learner answered correctly/incorrectly to which word question. The method visualizes the areas a Figure 1: Usage of “haunt” predicted to be familiar to the learner. learner user may know by classifying each usage point in the visualization into known/not known to the learner. This classification is conducted in a personalized manner because learners’ language skills and specialized fields are different. The learner only needs to take a 30-minute vocabulary test for this purpose. Figure 1 shows an example visualization using our method. “To haunt” has two different meanings in English, the first being “to chase” and the other “to curse,” or to be af- fected by ghosts or misfortune. Each point shows the usage Figure 2: An example of concordancer. of the word in a corpus written by native speakers. The dif- ferences in point colors indicate whether they are predicted to be known to the learner. The right side of the figure, within the dotted curve, is predicted to be known to the learner. In Related Work this way, our method visualizes the semantic area the learner Explainable machine learning studies knows. While deep learning-based methods outperformed conven- Our contribution is as follows: tional machine learning methods such as support vector ma- chines (SVMs) in many tasks, parameters of deep learn- • For second language vocabulary learning, we propose ing methods are typically more difficult to interpret com- a novel supervised visualization model that captures pared to those in conventional models. To this end, in the word broadness via a personalized prediction of learner’s machine learning and artificial intelligence community, a knowledge of usages. number of methods have been proposed to extract expla- nations from trained machine-learning models, or training • As our visualization uses a vocabulary test result dataset models taking explainability into accounts (Ribeiro, Singh, as supervisions, learners can understand which usage of and Guestrin 2016; Koh and Liang 2017; Lundberg and Lee the inputted word is predicted to be known/not known to 2017; Ribeiro, Singh, and Guestrin 2018). him/her. Unlike previous methods that output automatic However, the purpose of these methods is to explain explanation of machine-learning models, our method is machine-learning models to help machine-learning engi- much more intuitive and novice-friendly for language neers and researchers in understanding the models. Obvi- learners in the sense that language learners do not need ously, second language learners are usually not machine- to know about machine learning models. learning engineers and researchers. Therefore, methods of these studies have different purposes, and it is difficult to • We evaluated our method in terms of predictive accuracy apply these methods to help their understanding of the mod- of vocabulary test result dataset and achieved better re- els. Language learners are typically even not interested in sults compared to baselines. the models. Rather, learners’ interests reside in understand- ing their current learning status and what they should learn to improve it. Hence, to meet learners’ needs, a model is de- Concordancer studies sirable for a learner to see his/her current learning status and While our proposed method is novel as a visualization, what he/she needs to learn in the near future. software tools that search the usages of an inputted word for educational purposes and display them itself are not Word Embedding Visualization Studies novel: such software is known as concordancers. Concor- dancers target learners, educators, and linguists as primary Word embedding techniques are techniques that have been users.
Recommended publications
  • Preparation and Exploitation of Bilingual Texts Dusko Vitas, Cvetana Krstev, Eric Laporte
    Preparation and exploitation of bilingual texts Dusko Vitas, Cvetana Krstev, Eric Laporte To cite this version: Dusko Vitas, Cvetana Krstev, Eric Laporte. Preparation and exploitation of bilingual texts. Lux Coreana, 2006, 1, pp.110-132. hal-00190958v2 HAL Id: hal-00190958 https://hal.archives-ouvertes.fr/hal-00190958v2 Submitted on 27 Nov 2007 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Preparation and exploitation of bilingual texts Duško Vitas Faculty of Mathematics Studentski trg 16, CS-11000 Belgrade, Serbia Cvetana Krstev Faculty of Philology Studentski trg 3, CS-11000 Belgrade, Serbia Éric Laporte Institut Gaspard-Monge, Université de Marne-la-Vallée 5, bd Descartes, 77454 Marne-la-Vallée CEDEX 2, France Introduction A bitext is a merged document composed of two versions of a given text, usually in two different languages. An aligned bitext is produced by an alignment tool or aligner, that automatically aligns or matches the versions of the same text, generally sentence by sentence. A multilingual aligned corpus or collection of aligned bitexts, when consulted with a search tool, can be extremely useful for translation, language teaching and the investigation of literary text (Veronis, 2000).
    [Show full text]
  • The Iafor European Conference Series 2014 Ece2014 Ecll2014 Ectc2014 Official Conference Proceedings ISSN: 2188-1138
    the iafor european conference series 2014 ece2014 ecll2014 ectc2014 Official Conference Proceedings ISSN: 2188-1138 “To Open Minds, To Educate Intelligence, To Inform Decisions” The International Academic Forum provides new perspectives to the thought-leaders and decision-makers of today and tomorrow by offering constructive environments for dialogue and interchange at the intersections of nation, culture, and discipline. Headquartered in Nagoya, Japan, and registered as a Non-Profit Organization 一般社( 団法人) , IAFOR is an independent think tank committed to the deeper understanding of contemporary geo-political transformation, particularly in the Asia Pacific Region. INTERNATIONAL INTERCULTURAL INTERDISCIPLINARY iafor The Executive Council of the International Advisory Board IAB Chair: Professor Stuart D.B. Picken IAB Vice-Chair: Professor Jerry Platt Mr Mitsumasa Aoyama Professor June Henton Professor Frank S. Ravitch Director, The Yufuku Gallery, Tokyo, Japan Dean, College of Human Sciences, Auburn University, Professor of Law & Walter H. Stowers Chair in Law USA and Religion, Michigan State University College of Law Professor David N Aspin Professor Emeritus and Former Dean of the Faculty of Professor Michael Hudson Professor Richard Roth Education, Monash University, Australia President of The Institute for the Study of Long-Term Senior Associate Dean, Medill School of Journalism, Visiting Fellow, St Edmund’s College, Cambridge Economic Trends (ISLET) Northwestern University, Qatar University, UK Distinguished Research Professor of Economics,
    [Show full text]
  • Kernerman Kdictionaries.Com/Kdn DICTIONARY News the European Network of E-Lexicography (Enel) Tanneke Schoonheim
    Number 22 ● July 2014 Kernerman kdictionaries.com/kdn DICTIONARY News The European Network of e-Lexicography (ENeL) Tanneke Schoonheim On October 11th 2013, the kick-off meeting of the European production and reception of dictionaries. The internet offers Network of e-Lexicography (ENeL) project took place in entirely new possibilities for developing and presenting Brussels. This meeting was the outcome of an idea ventilated dictionary information, such as with the integration of sound, a year and a half earlier, in March 2012 in Berlin, at the maps or video, and various novel ways of interacting with European Workshop on Future Standards in Lexicography. dictionary users. For editors of scholarly dictionaries the new The workshop participants then confirmed the imperative to medium is not only a source of inspiration, it also generates coordinate and harmonise research in the field of (electronic) new and serious challenges that demand cooperation and lexicography across Europe, namely to share expertise relating standardization on various levels: to standards, discuss new methodologies in lexicography that a. Through the internet scholarly dictionaries can potentially fully exploit the possibilities of the digital medium, reflect on reach large audiences. However, at present scholarly the pan-European nature of the languages of Europe and attain dictionaries providing reliable information are often not easy a wider audience. to find and are hard to decode for a non-academic audience; A proposal was written by a team of researchers from
    [Show full text]
  • TANGO: Bilingual Collocational Concordancer
    TANGO: Bilingual Collocational Concordancer Jia-Yan Jian Yu-Chia Chang Jason S. Chang Department of Computer Inst. of Information Department of Computer Science System and Applictaion Science National Tsing Hua National Tsing Hua National Tsing Hua University University University 101, Kuangfu Road, 101, Kuangfu Road, 101, Kuangfu Road, Hsinchu, Taiwan Hsinchu, Taiwan Hsinchu, Taiwan [email protected] [email protected] [email protected] du.tw on elaborated statistical calculation. Moreover, log Abstract likelihood ratios are regarded as a more effective In this paper, we describe TANGO as a method to identify collocations especially when the collocational concordancer for looking up occurrence count is very low (Dunning, 1993). collocations. The system was designed to Smadja’s XTRACT is the pioneering work on answer user’s query of bilingual collocational extracting collocation types. XTRACT employed usage for nouns, verbs and adjectives. We first three different statistical measures related to how obtained collocations from the large associated a pair to be collocation type. It is monolingual British National Corpus (BNC). complicated to set different thresholds for each Subsequently, we identified collocation statistical measure. We decided to research and instances and translation counterparts in the develop a new and simple method to extract bilingual corpus such as Sinorama Parallel monolingual collocations. Corpus (SPC) by exploiting the word- We also provide a web-based user interface alignment technique. The main goal of the capable of searching those collocations and its concordancer is to provide the user with a usage. The concordancer supports language reference tools for correct collocation use so learners to acquire the usage of collocation.
    [Show full text]
  • Aconcorde: Towards a Proper Concordance for Arabic
    aConCorde: towards a proper concordance for Arabic Andrew Roberts, Dr Latifa Al-Sulaiti and Eric Atwell School of Computing University of Leeds LS2 9JT United Kingdom {andyr,latifa,eric}@comp.leeds.ac.uk July 12, 2005 Abstract Arabic corpus linguistics is currently enjoying a surge in activity. As the growth in the number of available Arabic corpora continues, there is an increased need for robust tools that can process this data, whether it be for research or teaching. One such tool that is useful for both groups is the concordancer — a simple tool for displaying a specified target word in its context. However, obtaining one that can reliably cope with the Arabic lan- guage had proved extremely difficult. Therefore, aConCorde was created to provide such a tool to the community. 1 Introduction A concordancer is a simple tool for summarising the contents of corpora based on words of interest to the user. Otherwise, manually navigating through (of- ten very large) corpora would be a long and tedious task. Concordancers are therefore extremely useful as a time-saving tool, but also much more. By isolat- ing keywords in their contexts, linguists can use concordance output to under- stand the behaviour of interesting words. Lexicographers can use the evidence to find if a word has multiple senses, and also towards defining their meaning. There is also much research and discussion about how concordance tools can be beneficial for data-driven language learning (Johns, 1990). A number of studies have demonstrated that providing access to corpora and concordancers benefited students learning a second language.
    [Show full text]
  • Panacea D6.1
    SEVENTH FRAMEWORK PROGRAMME THEME 3 Information and communication Technologies PANACEA Project Grant Agreement no.: 248064 Platform for Automatic, Normalized Annotation and Cost-Effective Acquisition of Language Resources for Human Language Technologies D6.1 Technologies and Tools for Lexical Acquisition Dissemination Level: Public Delivery Date: July 16th 2010 Status – Version: Final Author(s) and Affiliation: Laura Rimell (UCAM), Anna Korhonen (UCAM), Valeria Quochi (ILC-CNR), Núria Bel (UPF), Tommaso Caselli (ILC-CNR), Prokopis Prokopidis (ILSP), Maria Gavrilidou (ILSP), Thierry Poibeau (UCAM), Muntsa Padró (UPF), Eva Revilla (UPF), Monica Monachini(CNR-ILC), Maurizio Tesconi (CNR-IIT), Matteo Abrate (CNR-IIT) and Clara Bacciu (CNR-IIT) D6.1 Technologies and Tools for Lexical Acquisition This document is part of technical documentation generated in the PANACEA Project, Platform for Automatic, Normalized Annotation and Cost-Effective Acquisition (Grant Agreement no. 248064). This documented is licensed under a Creative Commons Attribution 3.0 Spain License. To view a copy of this license, visit http://creativecommons.org/licenses/by/3.0/es/. Please send feedback and questions on this document to: [email protected] TRL Group (Tecnologies dels Recursos Lingüístics), Institut Universitari de Lingüística Aplicada, Universitat Pompeu Fabra (IULA-UPF) D6.1 – Technologies and Tools for Lexical Acquisition Table of contents Table of contents ..........................................................................................................................
    [Show full text]
  • Translate's Localization Guide
    Translate’s Localization Guide Release 0.9.0 Translate Jun 26, 2020 Contents 1 Localisation Guide 1 2 Glossary 191 3 Language Information 195 i ii CHAPTER 1 Localisation Guide The general aim of this document is not to replace other well written works but to draw them together. So for instance the section on projects contains information that should help you get started and point you to the documents that are often hard to find. The section of translation should provide a general enough overview of common mistakes and pitfalls. We have found the localisation community very fragmented and hope that through this document we can bring people together and unify information that is out there but in many many different places. The one section that we feel is unique is the guide to developers – they make assumptions about localisation without fully understanding the implications, we complain but honestly there is not one place that can help give a developer and overview of what is needed from them, we hope that the developer section goes a long way to solving that issue. 1.1 Purpose The purpose of this document is to provide one reference for localisers. You will find lots of information on localising and packaging on the web but not a single resource that can guide you. Most of the information is also domain specific ie it addresses KDE, Mozilla, etc. We hope that this is more general. This document also goes beyond the technical aspects of localisation which seems to be the domain of other lo- calisation documents.
    [Show full text]
  • Corpus Linguistics and Data-Driven Learning: a Critical Overview
    Published in Bulletin VALS-ASLA 97, 97-118, 2013 which should be used for any reference to this work Corpus linguistics and data-driven learning: a critical overview Alex BOULTON Université de Lorraine et Crapel, ATILF (UMR 7118) 44 avenue de la libération, 54000 Nancy, France [email protected] Henry TYNE Université de Perpignan Via Domitia et VECT (EA 2983) 52 avenue Paul Alduy, 66000 Perpignan, France [email protected] L'utilisation de corpus s'est avérée intéressante dans de nombreux domaines depuis plus de vingt ans, notamment en didactique des langues. Un état de l'art des études publiées dans ce domaine nous permettra de constater que si les résultats sont en général assez positifs (l'apprentissage a lieu), l'exploitation de corpus reste une activité marginale. Si, en général, de telles recherches en didactique des langues ne donnent pas toujours des résultats fulgurants, on pourrait faire l’hypothèse que les approches sont plus ou moins adaptées à certains types d'apprenants, ou à certains types d'activité, et qu'elles contribuent au bagage général des outils et des techniques disponibles. Qui plus est, avec une approche sur corpus, on pourrait de même croire que les étudiants deviennent tout simplement des apprenants plus conscients de leur apprentissage grâce au travail effectué. Nous suggérons que l'apport des corpus pour l'apprentissage des langues offre ainsi de nombreuses possibilités. Mots-clés: Corpus, apprentissage sur corpus, pratiques ordinaires d'enseignement-apprentissage, processus, TIC 1. Introduction Corpus linguistics is not simply a recondite field of research within linguistics: it affords practical methodologies and tools to further the study of all aspects of language use (e.g.
    [Show full text]
  • From a Bilingual Concordancer to a Translation Finder Julien Bourdaillet, Stéphane Huet, Philippe Langlais, Guy Lapalme
    TransSearch: from a bilingual concordancer to a translation finder Julien Bourdaillet, Stéphane Huet, Philippe Langlais, Guy Lapalme To cite this version: Julien Bourdaillet, Stéphane Huet, Philippe Langlais, Guy Lapalme. TransSearch: from a bilingual concordancer to a translation finder. Machine Translation, Springer Verlag, 2010, 24 (3-4), pp.241-271. 10.1007/s10590-011-9089-6. hal-02021930 HAL Id: hal-02021930 https://hal.archives-ouvertes.fr/hal-02021930 Submitted on 26 Feb 2019 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. 2 J. Bourdaillet, S. Huet, P. Langlais, G. Lapalme Figure 1. English sentences, along with their French translation, where the query in keeping with occurs in the current version of the bilingual concordancer Trans- Search. The query is highlighted in the English sentences, but the user has to search its translations in the French sentences. This makes the discovery of the different translation possibilities more difficult (adapt´ees`a, conforme aux, d´etonne and conform´ement`a for the four displayed sentences), and forces the user to scroll down the page in order to read other sentences while searching for further translation material. This paper describes the enhancement of the web-based bilingual concordancer TransSearch1 (Macklovitch et al., 2000).
    [Show full text]
  • Computer Aided Translation Advances and Challenges
    Computer Aided Translation Advances and Challenges Philipp Koehn 7 August 2016 Philipp Koehn Computer Aided Translation 7 August 2016 Overview 1 • A practical introduction: the casmacat workbench • Postediting • Types of assistance • Logging, eye tracking and user studies • Implementation details of the casmacat workbench : Philipp Koehn Computer Aided Translation 7 August 2016 2 part I casmacat workbench Philipp Koehn Computer Aided Translation 7 August 2016 3 casmacat workbench Philipp Koehn Computer Aided Translation 7 August 2016 CASMACAT Project 2011-2014 4 • Cognitive studies of translators leading to insights into interface design ! better understanding of translator needs • Workbench with novel types of assistance to human translators { interactive translation prediction { interactive editing and reviewing { adaptive translation models ! better tools for translators • Demonstration of effectiveness in field tests with professional translators ! increased translator productivity Philipp Koehn Computer Aided Translation 7 August 2016 Core Modes 5 Philipp Koehn Computer Aided Translation 7 August 2016 Postediting Modes 6 Philipp Koehn Computer Aided Translation 7 August 2016 GUI Features 7 Philipp Koehn Computer Aided Translation 7 August 2016 Postediting Interface 8 • Source on left, translation on right • Context above and below Philipp Koehn Computer Aided Translation 7 August 2016 Confidence Measures 9 • Sentence-level confidence measures ! estimate usefulness of machine translation output • Word-level confidence measures ! point
    [Show full text]
  • The Universal Dependencies Treebank of Spoken Slovenian
    The Universal Dependencies Treebank of Spoken Slovenian Kaja Dobrovoljc1, Joakim Nivre2 1Institute for Applied Slovene Studies Trojina, Ljubljana, Slovenia 1Department of Slovenian Studies, Faculty of Arts, University of Ljubljana 2Department of Linguistics and Philology, Uppsala University [email protected], joakim.nivre@lingfil.uu.se Abstract This paper presents the construction of an open-source dependency treebank of spoken Slovenian, the first syntactically annotated collection of spontaneous speech in Slovenian. The treebank has been manually annotated using the Universal Dependencies annotation scheme, a one-layer syntactic annotation scheme with a high degree of cross-modality, cross-framework and cross-language interoper- ability. In this original application of the scheme to spoken language transcripts, we address a wide spectrum of syntactic particularities in speech, either by extending the scope of application of existing universal labels or by proposing new speech-specific extensions. The initial analysis of the resulting treebank and its comparison with the written Slovenian UD treebank confirms significant syntactic differences between the two language modalities, with spoken data consisting of shorter and more elliptic sentences, less and simpler nominal phrases, and more relations marking disfluencies, interaction, deixis and modality. Keywords: dependency treebank, spontaneous speech, Universal Dependencies 1. Introduction actually out-performs state-of-the-art pipeline approaches (Rasooli and Tetreault, 2013; Honnibal and Johnson, 2014). It is nowadays a well-established fact that data-driven pars- Such heterogeneity of spoken language annotation schemes ing systems used in different speech-processing applica- inevitably leads to a restricted usage of existing spoken tions benefit from learning on annotated spoken data, rather language treebanks in linguistic research and parsing sys- than using models built on written language observation.
    [Show full text]
  • Arxiv:2105.02877V1 [Cs.CV] 6 May 2021 Tles Corresponding to the Audio Content
    Aligning Subtitles in Sign Language Videos Hannah Bull1* Triantafyllos Afouras2∗ Gul¨ Varol2;3 Samuel Albanie2 Liliane Momeni2 Andrew Zisserman2 1 LISN, Univ Paris-Saclay, CNRS, France 2 Visual Geometry Group, University of Oxford, UK 3 LIGM, Ecole´ des Ponts, Univ Gustave Eiffel, CNRS, France [email protected]; fafourast,gul,albanie,liliane,[email protected] https://www.robots.ox.ac.uk/˜vgg/research/bslalign/ Overlooked by a small hill known as Leopard the mother must make sure To keep her cubs alive in this dangerous neighbourhood, Saudio Rock. they stay hidden. Sgt Overlooked by a small hill known as Leopard Rock. the mother must make sure they stay hidden. To keep her cubs alive in this dangerous neighbourhood, Time 14:12 14:14 14:16 14:18 14:20 14:22 14:24 14:26 14:28 14:30 Figure 1: Subtitle alignment: We study the task of aligning subtitles to continuous signing in sign language interpreted TV broadcast data. The subtitles in such settings usually correspond to and are aligned with the audio content (top: audio subtitles, Saudio) but are unaligned with the accompanying signing (bottom: Ground Truth annotation of the signing corresponding to the subtitle, Sgt). This is a very challenging task as (i) the order of subtitles varies between spoken and sign languages, (ii) the duration of a subtitle differs considerably between signing and speech, and (iii) the signing corresponds to a translation of the speech as opposed to a transcription. Abstract evaluations, we show substantial improvements over exist- ing alignment baselines that do not make use of subtitle text The goal of this work is to temporally align asyn- embeddings for learning.
    [Show full text]