Modeling and Encoding Traditional Wordlists for Machine Applications

Total Page:16

File Type:pdf, Size:1020Kb

Modeling and Encoding Traditional Wordlists for Machine Applications Modeling and Encoding Traditional Wordlists for Machine Applications Shakthi Poornima Jeff Good Department of Linguistics Department of Linguistics University at Buffalo University at Buffalo Buffalo, NY USA Buffalo, NY USA [email protected] [email protected] Abstract Clearly, descriptive linguistic resources can be This paper describes work being done on of potential value not just to traditional linguis- the modeling and encoding of a legacy re- tics, but also to computational linguistics. The source, the traditional descriptive wordlist, difficulty, however, is that the kinds of resources in ways that make its data accessible to produced in the course of linguistic description NLP applications. We describe an abstract are typically not easily exploitable in NLP appli- model for traditional wordlist entries and cations. Nevertheless, in the last decade or so, then provide an instantiation of the model it has become widely recognized that the devel- in RDF/XML which makes clear the re- opment of new digital methods for encoding lan- lationship between our wordlist database guage data can, in principle, not only help descrip- and interlingua approaches aimed towards tive linguists to work more effectively but also al- machine translation, and which also al- low them, with relatively little extra effort, to pro- lows for straightforward interoperation duce resources which can be straightforwardly re- with data from full lexicons. purposed for, among other things, NLP (Simons et al., 2004; Farrar and Lewis, 2007). 1 Introduction Despite this, it has proven difficult to create When looking at the relationship between NLP significant electronic descriptive resources due to and linguistics, it is typical to focus on the dif- the complex and specific problems inevitably as- ferent approaches taken with respect to issues sociated with the conversion of legacy data. One like parsing and generation of natural language exception to this is found in the work done in data—for example, to compare statistical NLP ap- the context of the ODIN project (Xia and Lewis, proaches to those involving grammar engineering. 2009), a significant database of interlinear glossed Such comparison is undoubtedly important insofar text (IGT), a standard descriptive linguistic data as it helps us understand how computational meth- format (Palmer et al., 2009), compiled by search- ods that are derived from these two lines of re- ing the Web for legacy instances of IGT. search can complement each other. However, one This paper describes another attempt to trans- thing that the two areas of work have in common form an existing legacy dataset into a more read- is that they tend to focus on majority languages ily repurposable format. Our data consists of tra- and majority language resources. Even where this ditional descriptive wordlists originally collected 1 is not the case (Bender et al., 2002; Alvarez et al., for comparative and historical linguistic research. 2006; Palmer et al., 2009), the resulting products Wordlists have been widely employed as a first still cover relatively few languages from a world- step towards the creation of a dictionary or as a wide perspective. This is in part because such means to quickly gather information about a lan- work cannot easily make use of the extensive lan- guage for the purposes of language comparison guage resources produced by descriptive linguists, (especially in parts of the world where languages the group of researchers that are most actively in- 1These wordlists were collected by Timothy Usher and volved in documenting the world’s entire linguis- Paul Whitehouse and represent an enormous effort without tic diversity. In fact, one particular descriptive lin- which the work described here would not have been possible. The RDF/XML implementations discussed in this paper will guistic product, the wordlist—which is the focus be made available at http://lego.linguistlist.org of this paper—can be found for at least a quarter within the context of the Lexicon Enhancement via the of the world’s languages. GOLD Ontology project. 1 Proceedings of the 2010 Workshop on NLP and Linguistics: Finding the Common Ground, ACL 2010, pages 1–9, Uppsala, Sweden, 16 July 2010. c 2010 Association for Computational Linguistics are poorly documented). Because of this, they 3 Related Work in Descriptive exist for many more languages than do full lexi- Linguistics cons. While the lexical information that wordlists Recent years have seen a fair amount of attention contain is quite sparse, they are relatively consis- paid to the modeling of traditional linguistic data tent in their structure across resources. This al- types, including lexicons, glossed texts, and gram- lows for the creation of a large-scale multilingual mars (Bell and Bird, 2000; Good, 2004; Palmer database consisting of rough translational equiva- and Erk, 2007; Nordhoff, 2008). The data type of lents which may lack precision but has coverage focus here, wordlists, has not seen serious treat- well-beyond what would otherwise be available. ment. Superficially, wordlists resemble lexicons 2 The Data and Project Background and, of course, they can be considered a kind of lexical resource. However, as will be shown in The data we are working with consists of 2,700 section 5, there are important differences between wordlists drawn from more than 1,500 languages lexicons and wordlists which have implications for (some wordlists represent dialects) and close to how they should be modeled. 500,000 forms. This is almost certainly the largest Most of the work on modeling descriptive lin- collection of wordlists in a standardized format. guistic data types has proceeded without special The average size of the individual wordlists is consideration for possible NLP applications for rather small, around 200 words, making them the data being encoded. This is largely because the comparable in size to the resources found in work was initially a response to issues relating to a project like NEDO (Takenobu, 2006), though the longevity of digital descriptive data which was, smaller than in other related projects like those otherwise, quite often being encoded solely in (of- discussed in section 4. While the work described ten proprietary) presentation formats (Bird and Si- here was originally conceived to support descrip- mons, 2003). However, the possibility for fruitful tive and comparative linguistics, our data model interaction between computational linguistics and and choice of encoding technologies has had the descriptive linguistics is apparent and has been the additional effect of making these resources read- subject of some work (Palmer et al., 2009). ily exploitable in other domains, in particular NLP. The work described here is also interested in We have approached the data initially as tradi- this possibility. In particular, we address the ques- tional, not computational, linguists, and our first tion of how to model and encode a large-scale goal has been to encode the available materials dataset that was originally intended to be used for not with any new information but rather to trans- descriptive purposes in ways that not only allow us fer the information they originally contained in a to faithfully represent the intention of the original more exploitable way. creator but also permit the data to be straightfor- By way of introduction, the hypothetical exam- wardly exploitable for new uses, including NLP. ple in (1) illustrates a traditional presentation for- To the best of our knowledge, our work is innova- mat of a wordlist, with English as the source lan- tive both because of the data type being explored guage and French as the target language. and because the data modeling is being done par- allel with the transformation of a legacy resource (1) MAN homme with significant coverage of the world’s languages. WOMAN femme This stands in contrast to most other work (again, As we will describe in more detail in section 5, with the exception of work done within ODIN they key features of a wordlist entry are an index (Xia and Lewis, 2009)) whose data, while repre- to a concept assumed to be of general provenance sentative, is not of the same scale. (e.g., MAN) and a form drawn from a specific lan- guage (e.g. homme) determined to be the counter- 4 Related Work on Lexicon part for that concept within that language. Most Interoperability in NLP typically, the elements indexing the relevant con- The relevant related work in NLP is that focused cepts are words drawn from languages of wider on interoperation among lexical resources. One communication (e.g., English or Spanish). way to achieve this is to make use of language in- dependent ontologies (or comparable objects) for word meanings which can serve as pivots for mul- 2 tilingual applications (Ide et al., 1998; Vossen, one begins with an abstract meaning, for example 2004; Nirenburg et al., 2004; Ronzano et al., MAN, and then tries to find the word in the tar- 2010). The word senses provided by WordNet, for get language which represents the best semantic example, have been used for this purpose (O’Hara fit for that meaning. Lexicons, on the other hand, et al., 1998). prototypically map in the opposite direction from A recognized data modeling standard for lexi- form to meaning. Furthermore, as will be elab- cal interoperation is the Lexical Markup Frame- orated in section 5.3, the meanings employed in work (LMF), which provides standardized frame- wordlists are not intended to refer to meanings of work for the description and representation of lex- lexical items in specific languages. In this way, icons (Francopoulo et al., 2009). Instantiations of they are quite distinct from bilingual dictionaries. LMF have also been extended to represent Word- We can therefore view a wordlist as a set of de- Nets, e.g., Wordnet-LMF (Soria et al., 2009), in fective signs—containing information on the form ways which facilitate interoperation.
Recommended publications
  • Automatic Wordnet Mapping Using Word Sense Disambiguation*
    Automatic WordNet mapping using word sense disambiguation* Changki Lee Seo JungYun Geunbae Leer Natural Language Processing Lab Natural Language Processing Lab Dept. of Computer Science and Engineering Dept. of Computer Science Pohang University of Science & Technology Sogang University San 31, Hyoja-Dong, Pohang, 790-784, Korea Sinsu-dong 1, Mapo-gu, Seoul, Korea {leeck,gblee }@postech.ac.kr seojy@ ccs.sogang.ac.kr bilingual Korean-English dictionary. The first Abstract sense of 'gwan-mog' has 'bush' as a translation in English and 'bush' has five synsets in This paper presents the automatic WordNet. Therefore the first sense of construction of a Korean WordNet from 'gwan-mog" has five candidate synsets. pre-existing lexical resources. A set of Somehow we decide a synset {shrub, bush} automatic WSD techniques is described for among five candidate synsets and link the sense linking Korean words collected from a of 'gwan-mog' to this synset. bilingual MRD to English WordNet synsets. As seen from this example, when we link the We will show how individual linking senses of Korean words to WordNet synsets, provided by each WSD method is then there are semantic ambiguities. To remove the combined to produce a Korean WordNet for ambiguities we develop new word sense nouns. disambiguation heuristics and automatic mapping method to construct Korean WordNet based on 1 Introduction the existing English WordNet. There is no doubt on the increasing This paper is organized as follows. In section 2, importance of using wide coverage ontologies we describe multiple heuristics for word sense for NLP tasks especially for information disambiguation for sense linking.
    [Show full text]
  • Concepticon: a Resource for the Linking of Concept Lists
    Concepticon: A Resource for the Linking of Concept Lists Johann-Mattis List1, Michael Cysouw2, Robert Forkel3 1CRLAO/EHESS and AIRE/UPMC, Paris, 2Forschungszentrum Deutscher Sprachatlas, Marburg, 3Max Planck Institute for the Science of Human History, Jena [email protected], [email protected], [email protected] Abstract We present an attempt to link the large amount of different concept lists which are used in the linguistic literature, ranging from Swadesh lists in historical linguistics to naming tests in clinical studies and psycholinguistics. This resource, our Concepticon, links 30 222 concept labels from 160 conceptlists to 2495 concept sets. Each concept set is given a unique identifier, a unique label, and a human-readable definition. Concept sets are further structured by defining different relations between the concepts. The resource can be used for various purposes. Serving as a rich reference for new and existing databases in diachronic and synchronic linguistics, it allows researchers a quick access to studies on semantic change, cross-linguistic polysemies, and semantic associations. Keywords: concepts, concept list, Swadesh list, naming test, word norms, cross-linguistically linked data 1. Introduction in the languages they were working on, or that they turned In 1950, Morris Swadesh (1909 – 1967) proposed the idea out to be not as stable and universal as Swadesh had claimed that certain parts of the lexicon of human languages are uni- (Matisoff, 1978; Alpher and Nash, 1999). Up to today, versal, stable over time, and rather resistant to borrowing. dozens of different concept lists have been compiled for var- As a result, he claimed that this part of the lexicon, which ious purposes.
    [Show full text]
  • Universal Or Variation? Semantic Networks in English and Chinese
    Universal or variation? Semantic networks in English and Chinese Understanding the structures of semantic networks can provide great insights into lexico- semantic knowledge representation. Previous work reveals small-world structure in English, the structure that has the following properties: short average path lengths between words and strong local clustering, with a scale-free distribution in which most nodes have few connections while a small number of nodes have many connections1. However, it is not clear whether such semantic network properties hold across human languages. In this study, we investigate the universal structures and cross-linguistic variations by comparing the semantic networks in English and Chinese. Network description To construct the Chinese and the English semantic networks, we used Chinese Open Wordnet2,3 and English WordNet4. The two wordnets have different word forms in Chinese and English but common word meanings. Word meanings are connected not only to word forms, but also to other word meanings if they form relations such as hypernyms and meronyms (Figure 1). 1. Cross-linguistic comparisons Analysis The large-scale structures of the Chinese and the English networks were measured with two key network metrics, small-worldness5 and scale-free distribution6. Results The two networks have similar size and both exhibit small-worldness (Table 1). However, the small-worldness is much greater in the Chinese network (σ = 213.35) than in the English network (σ = 83.15); this difference is primarily due to the higher average clustering coefficient (ACC) of the Chinese network. The scale-free distributions are similar across the two networks, as indicated by ANCOVA, F (1, 48) = 0.84, p = .37.
    [Show full text]
  • POS Tagging for Improving Code-Switching Identification In
    POS Tagging for Improving Code-Switching Identification in Arabic Mohammed Attia1 Younes Samih2 Ali Elkahky1 Hamdy Mubarak2 Ahmed Abdelali2 Kareem Darwish2 1Google LLC, New York City, USA 2Qatar Computing Research Institute, HBKU Research Complex, Doha, Qatar 1{attia,alielkahky}@google.com 2{ysamih,hmubarak,aabdelali,kdarwish}@hbku.edu.qa Abstract and sociolinguistic reasons. Second, existing the- oretical and computational linguistic models are When speakers code-switch between their na- tive language and a second language or lan- based on monolingual data and cannot adequately guage variant, they follow a syntactic pattern explain or deal with the influx of CS data whether where words and phrases from the embedded spoken or written. language are inserted into the matrix language. CS has been studied for over half a century This paper explores the possibility of utiliz- from different perspectives, including theoretical ing this pattern in improving code-switching linguistics (Muysken, 1995; Parkin, 1974), ap- identification between Modern Standard Ara- plied linguistics (Walsh, 1969; Boztepe, 2003; Se- bic (MSA) and Egyptian Arabic (EA). We try to answer the question of how strong is the tati, 1998), socio-linguistics (Barker, 1972; Heller, POS signal in word-level code-switching iden- 2010), psycho-linguistics (Grosjean, 1989; Prior tification. We build a deep learning model en- and Gollan, 2011; Kecskes, 2006), and more re- riched with linguistic features (including POS cently computational linguistics (Solorio and Liu, tags) that outperforms the state-of-the-art re- 2008a; Çetinoglu˘ et al., 2016; Adel et al., 2013b). sults by 1.9% on the development set and 1.0% In this paper, we investigate the possibility of on the test set.
    [Show full text]
  • Lexical Resource Reconciliation in the Xerox Linguistic Environment
    Lexical Resource Reconciliation in the Xerox Linguistic Environment Ronald M. Kaplan Paula S. Newman Xerox Palo Alto Research Center Xerox Palo Alto Research Center 3333 Coyote Hill Road 3333 Coyote Hill Road Palo Alto, CA, 94304, USA Palo Alto, CA, 94304, USA kapl an~parc, xerox, com pnewman©parc, xerox, tom Abstract This paper motivates and describes the morpho- logical and lexical adaptations of XLE. They evolved This paper motivates and describes those concurrently with PARGRAM, a multi-site XLF_~ aspects of the Xerox Linguistic Environ- based broad-coverage grammar writing effort aimed ment (XLE) that facilitate the construction at creating parallel grammars for English, French, of broad-coverage Lexical Functional gram- and German (see Butt et. al., forthcoming). The mars by incorporating morphological and XLE adaptations help to reconcile separately con- lexical material from external resources. structed linguistic resources with the needs of the Because that material can be incorrect, in- core grammars. complete, or otherwise incompatible with The paper is divided into three major sections. the grammar, mechanisms are provided to The next section sets the stage by providing a short correct and augment the external material overview of the overall environmental features of the to suit the needs of the grammar developer. original LFG GWB and its provisions for morpho- This can be accomplished without direct logical and lexicon processing. The two following modification of the incorporated material, sections describe the XLE extensions in those areas. which is often infeasible or undesirable. Externally-developed finite-state morpho- 2 The GWB Data Base logical analyzers are reconciled with gram- mar requirements by run-time simulation GWB provides a computational environment tai- of finite-state calculus operations for com- lored especially for defining and testing grammars bining transducers.
    [Show full text]
  • Wordnet As an Ontology for Generation Valerio Basile
    WordNet as an Ontology for Generation Valerio Basile To cite this version: Valerio Basile. WordNet as an Ontology for Generation. WebNLG 2015 1st International Workshop on Natural Language Generation from the Semantic Web, Jun 2015, Nancy, France. hal-01195793 HAL Id: hal-01195793 https://hal.inria.fr/hal-01195793 Submitted on 9 Sep 2015 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. WordNet as an Ontology for Generation Valerio Basile University of Groningen [email protected] Abstract a structured database of words in a format read- able by electronic calculators. For each word in In this paper we propose WordNet as an the database, WordNet provides a list of senses alternative to ontologies for the purpose of and their definition in plain English. The senses, natural language generation. In particu- besides having a inner identifier, are represented lar, the synset-based structure of WordNet as synsets, i.e., sets of synonym words. Words proves useful for the lexicalization of con- in general belong to multiple synsets, as they cepts, by providing ready lists of lemmas have more than one sense, so the relation between for each concept to generate.
    [Show full text]
  • 'Face' in Vietnamese
    Body part extensions with mặt ‘face’ in Vietnamese ​ Annika Tjuka 1 Introduction In many languages, body part terms have multiple meanings. The word tongue, for example, ​ ​ refers not only to the body part but also to a landscape or object feature, as in tongue of the ocean ​ or tongue of flame. These examples illustrate that body part terms are commonly mapped to ​ ​ concrete objects. In addition, body part terms can be transferred to abstract domains, including the description of emotional states (e.g., Yu 2002). The present study investigates the mapping of the body part term face to the concrete domain of objects. Expressions like mouth of the river ​ ​ and foot of the bed reveal the different analogies for extending human body parts to object ​ features. In the case of mouth of the river, the body part is extended on the basis of the function ​ ​ of the mouth as an opening. In contrast, the expression foot of the bed refers to the part of the bed ​ where your feet are if you lie in it. Thus, the body part is mapped according to the spatial alignment of the body part and the object. Ambiguous words challenge our understanding of how the mental lexicon is structured. Two processes could lead to the activation of a specific meaning: i) all meanings of a word are stored in one lexical entry and the relevant one is retrieved, ii) only a core meaning is stored and the other senses are generated by lexical rules (for a description of both models, see, Pustejovsky 1991). In the following qualitative analysis, I concentrate on the different meanings of the body part face and discuss the implications for its representation in the mental lexicon.
    [Show full text]
  • NL Assistant: a Toolkit for Developing Natural Language: Applications
    NL Assistant: A Toolkit for Developing Natural Language Applications Deborah A. Dahl, Lewis M. Norton, Ahmed Bouzid, and Li Li Unisys Corporation Introduction scale lexical resources have been integrated with the toolkit. These include Comlex and WordNet We will be demonstrating a toolkit for as well as additional, internally developed developing natural language-based applications resources. The second strategy is to provide easy and two applications. The goals of this toolkit to use editors for entering linguistic information. are to reduce development time and cost for natural language based applications by reducing Servers the amount of linguistic and programming work needed. Linguistic work has been reduced by Lexical information is supplied by four external integrating large-scale linguistics resources--- servers which are accessed by the natural Comlex (Grishman, et. al., 1993) and WordNet language engine during processing. Syntactic (Miller, 1990). Programming work is reduced by information is supplied by a lexical server based automating some of the programming tasks. The on the 50K word Comlex dictionary available toolkit is designed for both speech- and text- from the Linguistic Data Consortium. Semantic based interface applications. It runs in a information comes from two servers, a KB Windows NT environment. Applications can server based on the noun portion of WordNet run in either Windows NT or Unix. (70K concepts), and a semantics server containing case frame information for 2500 System Components English verbs. A denotations server using unique concept names generated for each The NL Assistant toolkit consists of WordNet synset at ISI links the words in the lexicon to the concepts in the KB.
    [Show full text]
  • Similarity Detection Using Latent Semantic Analysis Algorithm Priyanka R
    International Journal of Emerging Research in Management &Technology Research Article August ISSN: 2278-9359 (Volume-6, Issue-8) 2017 Similarity Detection Using Latent Semantic Analysis Algorithm Priyanka R. Patil Shital A. Patil PG Student, Department of Computer Engineering, Associate Professor, Department of Computer Engineering, North Maharashtra University, Jalgaon, North Maharashtra University, Jalgaon, Maharashtra, India Maharashtra, India Abstract— imilarity View is an application for visually comparing and exploring multiple models of text and collection of document. Friendbook finds ways of life of clients from client driven sensor information, measures the S closeness of ways of life amongst clients, and prescribes companions to clients if their ways of life have high likeness. Roused by demonstrate a clients day by day life as life records, from their ways of life are separated by utilizing the Latent Dirichlet Allocation Algorithm. Manual techniques can't be utilized for checking research papers, as the doled out commentator may have lacking learning in the exploration disciplines. For different subjective views, causing possible misinterpretations. An urgent need for an effective and feasible approach to check the submitted research papers with support of automated software. A method like text mining method come to solve the problem of automatically checking the research papers semantically. The proposed method to finding the proper similarity of text from the collection of documents by using Latent Dirichlet Allocation (LDA) algorithm and Latent Semantic Analysis (LSA) with synonym algorithm which is used to find synonyms of text index wise by using the English wordnet dictionary, another algorithm is LSA without synonym used to find the similarity of text based on index.
    [Show full text]
  • Introduction to Wordnet: an On-Line Lexical Database
    Introduction to WordNet: An On-line Lexical Database George A. Miller, Richard Beckwith, Christiane Fellbaum, Derek Gross, and Katherine Miller (Revised August 1993) WordNet is an on-line lexical reference system whose design is inspired by current psycholinguistic theories of human lexical memory. English nouns, verbs, and adjectives are organized into synonym sets, each representing one underlying lexical concept. Different relations link the synonym sets. Standard alphabetical procedures for organizing lexical information put together words that are spelled alike and scatter words with similar or related meanings haphazardly through the list. Unfortunately, there is no obvious alternative, no other simple way for lexicographers to keep track of what has been done or for readers to ®nd the word they are looking for. But a frequent objection to this solution is that ®nding things on an alphabetical list can be tedious and time-consuming. Many people who would like to refer to a dictionary decide not to bother with it because ®nding the information would interrupt their work and break their train of thought. In this age of computers, however, there is an answer to that complaint. One obvious reason to resort to on-line dictionariesÐlexical databases that can be read by computersÐis that computers can search such alphabetical lists much faster than people can. A dictionary entry can be available as soon as the target word is selected or typed into the keyboard. Moreover, since dictionaries are printed from tapes that are read by computers, it is a relatively simple matter to convert those tapes into the appropriate kind of lexical database.
    [Show full text]
  • Web Search Result Clustering with Babelnet
    Web Search Result Clustering with BabelNet Marek Kozlowski Maciej Kazula OPI-PIB OPI-PIB [email protected] [email protected] Abstract 2 Related Work In this paper we present well-known 2.1 Search Result Clustering search result clustering method enriched with BabelNet information. The goal is The goal of text clustering in information retrieval to verify how Babelnet/Babelfy can im- is to discover groups of semantically related docu- prove the clustering quality in the web ments. Contextual descriptions (snippets) of docu- search results domain. During the evalua- ments returned by a search engine are short, often tion we tested three algorithms (Bisecting incomplete, and highly biased toward the query, so K-Means, STC, Lingo). At the first stage, establishing a notion of proximity between docu- we performed experiments only with tex- ments is a challenging task that is called Search tual features coming from snippets. Next, Result Clustering (SRC). Search Results Cluster- we introduced new semantic features from ing (SRC) is a specific area of documents cluster- BabelNet (as disambiguated synsets, cate- ing. gories and glosses describing synsets, or Approaches to search result clustering can be semantic edges) in order to verify how classified as data-centric or description-centric they influence on the clustering quality (Carpineto, 2009). of the search result clustering. The al- The data-centric approach (as Bisecting K- gorithms were evaluated on AMBIENT Means) focuses more on the problem of data clus- dataset in terms of the clustering quality. tering, rather than presenting the results to the user. Other data-centric methods use hierarchical 1 Introduction agglomerative clustering (Maarek, 2000) that re- In the previous years, Web clustering engines places single terms with lexical affinities (2-grams (Carpineto, 2009) have been proposed as a solu- of words) as features, or exploit link information tion to the issue of lexical ambiguity in Informa- (Zhang, 2008).
    [Show full text]
  • SELECTING and CREATING a WORD LIST for ENGLISH LANGUAGE TEACHING by Deny A
    Teaching English with Technology, 17(1), 60-72, http://www.tewtjournal.org 60 SELECTING AND CREATING A WORD LIST FOR ENGLISH LANGUAGE TEACHING by Deny A. Kwary and Jurianto Universitas Airlangga Dharmawangsa Dalam Selatan, Surabaya 60286, Indonesia d.a.kwary @ unair.ac.id / [email protected] Abstract Since the introduction of the General Service List (GSL) in 1953, a number of studies have confirmed the significant role of a word list, particularly GSL, in helping ESL students learn English. Given the recent development in technology, several researchers have created word lists, each of them claims to provide a better coverage of a text and a significant role in helping students learn English. This article aims at analyzing the claims made by the existing word lists and proposing a method for selecting words and a creating a word list. The result of this study shows that there are differences in the coverage of the word lists due to the difference in the corpora and the source text analysed. This article also suggests that we should create our own word list, which is both personalized and comprehensive. This means that the word list is not just a list of words. The word list needs to be accompanied with the senses and the patterns of the words, in order to really help ESL students learn English. Keywords: English; GSL; NGSL; vocabulary; word list 1. Introduction A word list has been noted as an essential resource for language teaching, especially for second language teaching. The main purpose for the creation of a word list is to determine the words that the learners need to know, in order to provide more focused learning materials.
    [Show full text]