The LEGO Unified Concepticon
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Concepticon: a Resource for the Linking of Concept Lists
Concepticon: A Resource for the Linking of Concept Lists Johann-Mattis List1, Michael Cysouw2, Robert Forkel3 1CRLAO/EHESS and AIRE/UPMC, Paris, 2Forschungszentrum Deutscher Sprachatlas, Marburg, 3Max Planck Institute for the Science of Human History, Jena [email protected], [email protected], [email protected] Abstract We present an attempt to link the large amount of different concept lists which are used in the linguistic literature, ranging from Swadesh lists in historical linguistics to naming tests in clinical studies and psycholinguistics. This resource, our Concepticon, links 30 222 concept labels from 160 conceptlists to 2495 concept sets. Each concept set is given a unique identifier, a unique label, and a human-readable definition. Concept sets are further structured by defining different relations between the concepts. The resource can be used for various purposes. Serving as a rich reference for new and existing databases in diachronic and synchronic linguistics, it allows researchers a quick access to studies on semantic change, cross-linguistic polysemies, and semantic associations. Keywords: concepts, concept list, Swadesh list, naming test, word norms, cross-linguistically linked data 1. Introduction in the languages they were working on, or that they turned In 1950, Morris Swadesh (1909 – 1967) proposed the idea out to be not as stable and universal as Swadesh had claimed that certain parts of the lexicon of human languages are uni- (Matisoff, 1978; Alpher and Nash, 1999). Up to today, versal, stable over time, and rather resistant to borrowing. dozens of different concept lists have been compiled for var- As a result, he claimed that this part of the lexicon, which ious purposes. -
POS Tagging for Improving Code-Switching Identification In
POS Tagging for Improving Code-Switching Identification in Arabic Mohammed Attia1 Younes Samih2 Ali Elkahky1 Hamdy Mubarak2 Ahmed Abdelali2 Kareem Darwish2 1Google LLC, New York City, USA 2Qatar Computing Research Institute, HBKU Research Complex, Doha, Qatar 1{attia,alielkahky}@google.com 2{ysamih,hmubarak,aabdelali,kdarwish}@hbku.edu.qa Abstract and sociolinguistic reasons. Second, existing the- oretical and computational linguistic models are When speakers code-switch between their na- tive language and a second language or lan- based on monolingual data and cannot adequately guage variant, they follow a syntactic pattern explain or deal with the influx of CS data whether where words and phrases from the embedded spoken or written. language are inserted into the matrix language. CS has been studied for over half a century This paper explores the possibility of utiliz- from different perspectives, including theoretical ing this pattern in improving code-switching linguistics (Muysken, 1995; Parkin, 1974), ap- identification between Modern Standard Ara- plied linguistics (Walsh, 1969; Boztepe, 2003; Se- bic (MSA) and Egyptian Arabic (EA). We try to answer the question of how strong is the tati, 1998), socio-linguistics (Barker, 1972; Heller, POS signal in word-level code-switching iden- 2010), psycho-linguistics (Grosjean, 1989; Prior tification. We build a deep learning model en- and Gollan, 2011; Kecskes, 2006), and more re- riched with linguistic features (including POS cently computational linguistics (Solorio and Liu, tags) that outperforms the state-of-the-art re- 2008a; Çetinoglu˘ et al., 2016; Adel et al., 2013b). sults by 1.9% on the development set and 1.0% In this paper, we investigate the possibility of on the test set. -
'Face' in Vietnamese
Body part extensions with mặt ‘face’ in Vietnamese Annika Tjuka 1 Introduction In many languages, body part terms have multiple meanings. The word tongue, for example, refers not only to the body part but also to a landscape or object feature, as in tongue of the ocean or tongue of flame. These examples illustrate that body part terms are commonly mapped to concrete objects. In addition, body part terms can be transferred to abstract domains, including the description of emotional states (e.g., Yu 2002). The present study investigates the mapping of the body part term face to the concrete domain of objects. Expressions like mouth of the river and foot of the bed reveal the different analogies for extending human body parts to object features. In the case of mouth of the river, the body part is extended on the basis of the function of the mouth as an opening. In contrast, the expression foot of the bed refers to the part of the bed where your feet are if you lie in it. Thus, the body part is mapped according to the spatial alignment of the body part and the object. Ambiguous words challenge our understanding of how the mental lexicon is structured. Two processes could lead to the activation of a specific meaning: i) all meanings of a word are stored in one lexical entry and the relevant one is retrieved, ii) only a core meaning is stored and the other senses are generated by lexical rules (for a description of both models, see, Pustejovsky 1991). In the following qualitative analysis, I concentrate on the different meanings of the body part face and discuss the implications for its representation in the mental lexicon. -
SELECTING and CREATING a WORD LIST for ENGLISH LANGUAGE TEACHING by Deny A
Teaching English with Technology, 17(1), 60-72, http://www.tewtjournal.org 60 SELECTING AND CREATING A WORD LIST FOR ENGLISH LANGUAGE TEACHING by Deny A. Kwary and Jurianto Universitas Airlangga Dharmawangsa Dalam Selatan, Surabaya 60286, Indonesia d.a.kwary @ unair.ac.id / [email protected] Abstract Since the introduction of the General Service List (GSL) in 1953, a number of studies have confirmed the significant role of a word list, particularly GSL, in helping ESL students learn English. Given the recent development in technology, several researchers have created word lists, each of them claims to provide a better coverage of a text and a significant role in helping students learn English. This article aims at analyzing the claims made by the existing word lists and proposing a method for selecting words and a creating a word list. The result of this study shows that there are differences in the coverage of the word lists due to the difference in the corpora and the source text analysed. This article also suggests that we should create our own word list, which is both personalized and comprehensive. This means that the word list is not just a list of words. The word list needs to be accompanied with the senses and the patterns of the words, in order to really help ESL students learn English. Keywords: English; GSL; NGSL; vocabulary; word list 1. Introduction A word list has been noted as an essential resource for language teaching, especially for second language teaching. The main purpose for the creation of a word list is to determine the words that the learners need to know, in order to provide more focused learning materials. -
Modeling and Encoding Traditional Wordlists for Machine Applications
Modeling and Encoding Traditional Wordlists for Machine Applications Shakthi Poornima Jeff Good Department of Linguistics Department of Linguistics University at Buffalo University at Buffalo Buffalo, NY USA Buffalo, NY USA [email protected] [email protected] Abstract Clearly, descriptive linguistic resources can be This paper describes work being done on of potential value not just to traditional linguis- the modeling and encoding of a legacy re- tics, but also to computational linguistics. The source, the traditional descriptive wordlist, difficulty, however, is that the kinds of resources in ways that make its data accessible to produced in the course of linguistic description NLP applications. We describe an abstract are typically not easily exploitable in NLP appli- model for traditional wordlist entries and cations. Nevertheless, in the last decade or so, then provide an instantiation of the model it has become widely recognized that the devel- in RDF/XML which makes clear the re- opment of new digital methods for encoding lan- lationship between our wordlist database guage data can, in principle, not only help descrip- and interlingua approaches aimed towards tive linguists to work more effectively but also al- machine translation, and which also al- low them, with relatively little extra effort, to pro- lows for straightforward interoperation duce resources which can be straightforwardly re- with data from full lexicons. purposed for, among other things, NLP (Simons et al., 2004; Farrar and Lewis, 2007). 1 Introduction Despite this, it has proven difficult to create When looking at the relationship between NLP significant electronic descriptive resources due to and linguistics, it is typical to focus on the dif- the complex and specific problems inevitably as- ferent approaches taken with respect to issues sociated with the conversion of legacy data. -
TEI and the Documentation of Mixtepec-Mixtec Jack Bowers
Language Documentation and Standards in Digital Humanities: TEI and the documentation of Mixtepec-Mixtec Jack Bowers To cite this version: Jack Bowers. Language Documentation and Standards in Digital Humanities: TEI and the documen- tation of Mixtepec-Mixtec. Computation and Language [cs.CL]. École Pratique des Hauts Études, 2020. English. tel-03131936 HAL Id: tel-03131936 https://tel.archives-ouvertes.fr/tel-03131936 Submitted on 4 Feb 2021 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Préparée à l’École Pratique des Hautes Études Language Documentation and Standards in Digital Humanities: TEI and the documentation of Mixtepec-Mixtec Soutenue par Composition du jury : Jack BOWERS Guillaume, JACQUES le 8 octobre 2020 Directeur de Recherche, CNRS Président Alexis, MICHAUD Chargé de Recherche, CNRS Rapporteur École doctorale n° 472 Tomaž, ERJAVEC Senior Researcher, Jožef Stefan Institute Rapporteur École doctorale de l’École Pratique des Hautes Études Enrique, PALANCAR Directeur de Recherche, CNRS Examinateur Karlheinz, MOERTH Senior Researcher, Austrian Center for Digital Humanities Spécialité and Cultural Heritage Examinateur Linguistique Emmanuel, SCHANG Maître de Conférence, Université D’Orléans Examinateur Benoit, SAGOT Chargé de Recherche, Inria Examinateur Laurent, ROMARY Directeur de recherche, Inria Directeur de thèse 1. -
THE IMPACT of USING WORDLISTS in the LANGUAGE CLASSROOM on STUDENTS’ VOCABULARY ACQUISITION Gülçin Coşgun Ozyegin University
International Journal of English Language Teaching Vol.4, No.3, pp.49-66, March 2016 ___Published by European Centre for Research Training and Development UK (www.eajournals.org) THE IMPACT OF USING WORDLISTS IN THE LANGUAGE CLASSROOM ON STUDENTS’ VOCABULARY ACQUISITION Gülçin Coşgun Ozyegin University ABSTRACT: Vocabulary has always been an area of interest for many researchers since words represent “the building block upon which knowledge of the second language can be built” and without them people cannot convey the intended meaning (Abrudan, 2010). Nation (1988) emphasised if teachers want to help their learners to deal with unknown words, it would be better to spend more time on vocabulary learning strategies rather than spending time on individual words. However, Schmitt in Schmitt and McCarthy (1997:200) stated that among vocabulary learning strategies only ‘guessing from context’ and ‘key word method’ have been investigated in depth. Therefore, there is need for more research on vocabulary learning whose pedagogical implications may contribute to the field of second language learning. Considering the above- mentioned issues, vocabulary is a worthwhile field to investigate. Hence, this paper aims at proposing a framework for vocabulary teaching strategy in English as a foreign language context. KEYWORDS: Wordlists, Vocabulary Teaching, Word walls, Vocabulary Learning Strategies INTRODUCTION Vocabulary has always been an area of interest and concern for many researchers and teachers alike since words represent “the building block upon which knowledge of the second language can be built” and without them people cannot convey the intended meaning (Abrudan, 2010). Anderson and Freebody (1981) state that a strong correlation exists between vocabulary and academic achievement. -
1 from Text to Thought How Analyzing Language Can Advance
1 From Text to Thought How Analyzing Language Can Advance Psychological Science Joshua Conrad Jackson1*, Joseph Watts2,3,4, Johann-Mattis List2, Ryan Drabble1, Kristen Lindquist1 1Department of Psychology and Neuroscience, University of North Carolina at Chapel Hill 2Department of Linguistic and Cultural Evolution, Max Planck Institute for the Science of Human History 3Center for Research on Evolution, Belief, and Behaviour, University of Otago 4Religion Programme, University of Otago *Corresponding author Main Text Word Count, Excluding References: 8,116 Reference Count: 123 Acknowledgements: JCJ is supported by the National Science Foundation, the Royster Society of Fellows, and the John Templeton Foundation. JW is supported by the John Templeton Foundation and the Marsden Foundation of New Zealand (19-VUW-188). JML is supported by the European Research Council. The views in this paper do not necessarily reflect the views of these funding agencies. Abstract: Humans have been using language for thousands of years, but psychologists seldom consider what natural language can tell us about the mind. Here we propose that language offers a unique window into human cognition. After briefly summarizing the legacy of language analyses in psychological science, we show how methodological advances have made these analyses more feasible and insightful than ever before. In particular, we describe how two forms of language analysis—comparative linguistics and natural language processing—are already contributing to how we understand emotion, creativity, and religion, and overcoming methodological obstacles related to statistical power and culturally diverse samples. We summarize resources for learning both of these methods, and highlight the best way to combine language analysis techniques with behavioral paradigms. -
All-Words Word Sense Disambiguation Using Concept Embeddings
All-words Word Sense Disambiguation Using Concept Embeddings Rui Suzuki, Kanako Komiya, Masayuki Asahara, Minoru Sasaki, Hiroyuki Shinnou Ibaraki University, 4-12-1 Nakanarusawa, Hitachi, Ibaraki JAPAN, National Institute for Japanese Language and Linguistics, 10-2 Midoricho, Tachikawa, Tokyo, JAPAN, [email protected], kanako.komiya.nlp, minoru.sasaki.01, hiroyuki.shinnou.0828g @vc.ibaraki.ac.jp, [email protected] Abstract All-words word sense disambiguation (all-words WSD) is the task of identifying the senses of all words in a document. Since the sense of a word depends on the context, such as the surrounding words, similar words are believed to have similar sets of surrounding words. We therefore predict the target word senses by calculating the distances between the surrounding word vectors of the target words and their synonyms using word embeddings. In addition, we introduce the new idea of concept embeddings, constructed from concept tag sequences created from the results of previous prediction steps. We predict the target word senses using the distances between surrounding word vectors constructed from word and concept embeddings, via a bootstrapped iterative process. Experimental results show that these concept embeddings were able to improve the performance of Japanese all-words WSD. Keywords: word sense disambiguation, all-words, unsupervised 1. Introduction many words have the same article numbers. Several words Word sense disambiguation (WSD) involves identifying the can have the same precise article number, even when the senses of words in documents. In particular, the WSD task semantic breaks are considered. where the senses of all the words in a document are disam- biguated is referred to as all-words WSD. -
A Web-Based Vocabulary Profiler for Chinese Language Teaching and Research
DA A web-based vocabulary profiler for Chinese language teaching and research Jun Da, Middle Tennessee State University This paper describes the design and development of a web-based Chinese vocabulary profiler that is intended as a handy tool for Chinese language instruction and research. Given a piece of text, the profiler is capable of generating character and vocabulary profiles with frequency and availability information based on a comparison of the set of characters and words found in the text with a selection of other pre-compiled reference lists such as Da’s Modern Chinese character and N-gram frequency lists, the HSK word list and customized vocabulary list provided by a user. 1. Introduction Vocabulary profiling is a measure of the proportions of low and high frequency vocabulary used in a written text. In addition to frequency information, a profiler designed for lexical analysis of texts often provides other information such as the presence/absence of the set of words from the input text in other specialized word lists and their collocation with other words, etc. While it has been used in stylistic studies of written texts (e.g., Graves’s (2004) profiling of Jane Austen’s novels and letters), vocabulary profiling has also been suggested as a useful instrument in second language acquisition research and pedagogy (Laufer and Nation 1995). For example, Morris and Cobb (2004) examined the potential of using vocabulary profile as a predictor of TESL (Teaching English as a Second Language) trainees’ academic performance and reported that vocabulary profile results correlate significantly with their course grades. -
Proceedings of the 55Th Annual Meeting of the Association For
ComputEL-3 Proceedings of the 3rd Workshop on the Use of Computational Methods in the Study of Endangered Languages Volume 1 (Papers) February 26–27, 2019 Honolulu, Hawai‘i, USA Support: c 2019 The Association for Computational Linguistics Order copies of this and other ACL proceedings from: Association for Computational Linguistics (ACL) 209 N. Eighth Street Stroudsburg, PA 18360 USA Tel: +1-570-476-8006 Fax: +1-570-476-0860 [email protected] ISBN 978-1-950737-18-5 ii Preface These proceedings contain the papers presented at the 3rd Workshop on the Use of Computational Methods in the Study of Endangered languages held in Hawai’i at Manoa,¯ February 26–27, 2019. As the name implies, this is the third workshop held on the topic—the first meeting was co-located with the ACL main conference in Baltimore, Maryland in 2014 and the second one in 2017 was co-located with the 5th International Conference on Language Documentation and Conservation (ICLDC) at the University of Hawai‘i at Manoa.¯ The workshop covers a wide range of topics relevant to the study and documentation of endangered languages, ranging from technical papers on working systems and applications, to reports on community activities with supporting computational components. The purpose of the workshop is to bring together computational researchers, documentary linguists, and people involved with community efforts of language documentation and revitalization to take part in both formal and informal exchanges on how to integrate rapidly evolving language processing methods and tools into efforts of language description, documentation, and revitalization. The organizers are pleased with the range of papers, many of which highlight the importance of interdisciplinary work and interaction between the various communities that the workshop is aimed towards. -
Simple Features for Statistical Word Sense Disambiguation
SENSEVAL-3: Third International Workshop on the Evaluation of Systems for the Semantic Analysis of Text, Barcelona, Spain, July 2004 Association for Computational Linguistics Simple Features for Statistical Word Sense Disambiguation Abolfazl K. Lamjiri, Osama El Demerdash, Leila Kosseim CLaC Laboratory Department of Computer Science Concordia University, Montreal, Canada fa keigho,osama el,[email protected] Abstract more than four words around the target are In this paper, we describe our experiments on considered (Ide and V´eronis, 1998). While re- statistical word sense disambiguation (WSD) searchers such as Masterman (1961), Gougen- using two systems based on different ap- heim and Michea (1961), agree with this ob- proaches: Na¨ıve Bayes on word tokens and Max- servation (Ide and V´eronis, 1998), our results imum Entropy on local syntactic and seman- demonstrate that this does not generally ap- tic features. In the first approach, we consider ply to all words. A large context window pro- a context window and a sub-window within it vides domain information which increases the around the word to disambiguate. Within the accuracy for some target words such as bank.n, outside window, only content words are con- but not others like different.a or use.v (see Sec- sidered, but within the sub-window, all words tion 3). This confirms Mihalcea's observations are taken into account. Both window sizes are (Mihalcea, 2002). In our system we allow a tuned by the system for each word to disam- larger context window size and for most of the biguate and accuracies of 75% and 67% were re- words such context window is selected by the spectively obtained for coarse and fine grained system.