Towards a Computational Model of Grammaticalization and Lexical Diversity

Total Page:16

File Type:pdf, Size:1020Kb

Towards a Computational Model of Grammaticalization and Lexical Diversity Towards a computational model of grammaticalization and lexical diversity Christian Bentz Paula Buttery University of Cambridge, DTAL University of Cambridge, DTAL 9 West Road, CB3 9DA 9 West Road, CB3 9DA [email protected] [email protected] Abstract To model these processes Section 2 will Languages use different lexical inven- present a quantitative measure of lexical diver- tories to encode information, ranging sity based on Zipf-Mandelbrots law, which is from small sets of simplex words to also used as a biodiversity index (Jost, 2006). large sets of morphologically complex Based on this measure we present a prelimi- words. Grammaticalization theories nary computational model to reconstruct the argue that this variation arises as gradual change from lexically constrained to the outcome of diachronic processes lexically rich languages in Section 3. We whereby co-occurring words merge therefore use a simple grammaticalization al- to one word and build up complex gorithm and show how historical developments morphology. To model these pro- towards higher lexical diversity match the vari- cesses we present a) a quantitative ation in lexical diversity of natural languages measure of lexical diversity and b) a today. This suggests that synchronic variation preliminary computational model of in lexical diversity can be explained as the out- changes in lexical diversity over several come of diachronic language change. generations of merging higly frequent The computational model we present will collocates. therefore help to a) understand the diver- sity of lexical encoding strategies across lan- guages better, and b) to further uncover the 1 Introduction diachronic processes leading up to these syn- All languages share the property of being car- chronic differences. riers of information. However, they vastly dif- 2 Zipf’s law as a measure of lexical fer in terms of the exact encoding strategies diversity they adopt. For example, German encodes in- formation about number, gender, case, tense, Zipf-Mandelbrot’s law (Mandelbrot, 1953; aspect, etc. in a multitude of different articles, Zipf, 1949) states that ordering of words ac- pronouns, nouns, adjectives and verbs. This cording to their frequencies in texts will render abundant set of word forms contrasts with a frequency distributions of a specific shape: in smaller set of uninflected words in English. general, few words have high frequencies, fol- Crucially, grammaticalization theories lowed by a middle ground of medium frequen- (Heine and Kuteva, 2007, 2002; Bybee 2006, cies and a long tail of low frequency items. 2003; Hopper and Traugott, 2003; Lehmann, However, a series of studies pointed out that 1985) demonstrate that complex morpho- there are subtle differences in frequency dis- logical marking can derive diachronically by tributions for different texts and languages merging originally independent word forms (Bentz et al., forthcoming; Ha et al., 2006; that frequently co-occur. Over several gen- Popescu and Altmann, 2008). Namely, lan- erations of language learning and usage such guages with complex morphology tend to have grammaticalization and entrenchment pro- longer tails of low frequency words than lan- cesses can gradually increase the complexity guages with simplex morphology. The param- of word forms and hence the lexical diversity eters of Zipf-Mandelbrot’s law reflect these dif- of languages. ferences, and can be used as a quantitative 38 Proc. of 5th Workshop on Cognitive Aspects of Computational Language Learning (CogACLL) @ EACL 2014, pages 38–42, Gothenburg, Sweden, April 26 2014. c 2014 Association for Computational Linguistics measure of lexical diversity. 2.1 Method We use the definition of ZM’s law as captured by equation (1): C f(ri) = α , β + ri C > 0, α > 0, β > 1, i = 1, 2, . , n (1) − where f(ri) is the frequency of the word th of the i rank (ri), n is the number of ranks, C is a normalizing factor and α and β are parameters. To illustrate this, we use parallel Figure 1: Zipf frequency distributions for four texts of the Universal Declaration of Human natural languages (Fijian, English, German, Rights (UDHR) for Fijian, English, German Hungarian). Plots are in log-log space, val- and Hungarian. For frequency distributions ues 0.15, 0.1 and 0.05 were added to Fijian, of these texts (with tokens delimited by English and German log-frequencies to avoid white spaces) we can approximate the best overplotting. Values for the Zipf-Mandelbrot fitting parameters of the ZM law by means parameters are given in the legend. The of maximum likelihood estimation (Izs´ak, straight black line is the line of best fit for 2006; Murphy, 2013). In double logarithmic Fijian. space (see Figure 1) the normalizing factor C would shift the line of best fit upwards or low lexical diversity language like Fijian might downwards, α is the slope of this line and β develop towards a high diversity language like is Mandelbrot’s (1953) corrective for the fact Hungarian. that the line of best fit will deviate from a straight line for higher frequencies (upper left 3 Modelling changes in lexical corner in Figure 1). diversity As can be seen in Figure 1 Fijian has higher frequencies towards the lowest ranks (upper Grammaticalization theorists have long left corner) but the shortest tail of words with claimed that synchronic variation in word frequency one (horizontal bars in the lower complexity and lexical diversity might be the right corner). For Hungarian the pattern runs outcome of diachronic processes. Namely, the the other way round: it has the lowest frequen- grammaticalization cline from content item cies towards the low ranks and a long tail of >grammatical word >clitic >inflectional affix words with frequency one. German and En- is seen as a ubiquitous process in language glish lie between these. These patterns are re- change (Hopper and Traugott, 2003: 7). flected in ZM parameter values. Namely, Fi- In the final stage frequently co-occurring jian has the highest parameters, followed by words merge by means of phonological fusion English, German and Hungarian. By trend (Bybee, 2003: 617) and hence ’morphologize’ there is a negative relationship between ZM to built inflections and derivations. parameters and lexical diversity: low lexical Typical examples of a full cline of grammat- diversity is associated with high parameters, icalization are the Old English noun l¯ıc ’body’ high diversity is associated with low param- becoming the derivational suffix -ly, the inflec- eters. Cross-linguistically this effect can be tional future in Romance languages such as used to measure lexical diversity by means of Italian canter`o ’I will sing’ derived from Latin approximating the parameters of ZM’s law for cantare habeo ’I have to sing’, or Hungarian parallel texts. inflectional elative and inessive case markers In the following, we will present a compu- derived from a noun originally meaning ‘in- tational model to elicit the diachronic path- terior’ (Heine and Kuteva, 2007: 66). These ways of grammaticalization through which a processes can cause languages to distinguish 39 between a panoply of different word forms. For p : Percentage of words replaced by new • v example, Hungarian displays up to 20 different words. Choose pv of words randomly and noun forms where English would use a single replace all instances of these words by in- form (e.g. ship corresponding to Hungarian verting the letters. This simulates neolo- haj´o ’ship’, haj´oban ’in the ship’, haj´oba ’into gisms and loanwords replacing deprecated the ship’, etc.). words. As a consequence, once the full grammati- calization cline is completed this will increase rR: Range of ranks to be included in pv • the lexical diversity of a language. Note, replacements. If set to 0, vocabulary from however, that borrowings (loanwords) and ne- anywhere in the distribution will be ran- ologisms can also increase lexical diversity. domly replaced. Hence, a model of changes in lexical diversity n : Number of generations to simulate. will have to take both grammaticalization and • G new vocabulary into account. This simulation essentially allows us to vary 3.1 The model the degree of grammaticalization by means of Text: We use the Fijian UDHR as our start- varying pm, and also to control for the fact ing point for two reasons: a) Fijian is a lan- that frequency distributions might change due guage that is well known to be largely lack- to loanword borrowing and introduction of ing complex morphology, b) the UDHR is a new vocabulary (pv). Additionally, rR allows parallel text and hence allows us to compare us to vary the range of ranks where new words different languages by controlling for constant might replace deprecated ones. For frequency information content. Fijian has relatively low distributions calculated by generations we ap- lexical diversity and high ZM parameter val- proximate ZM parameters by maximum likeli- ues (see Figure 1). The question is whether hood estimations and therefore document the we can simulate a simple merging process over change of their shape. several generations that will transform the fre- Results: Figure 2 illustrates a simulation quency distribution of the original Fijian text of how the low lexical diversity language Fi- to fit the frequency distribution of the mor- jian approaches quantitative lexical properties phologically and lexically rich Hungarian text. similar to the Hungarian text just by means of To answer this question, we simulate the out- merging high-frequent collocates. While the come of grammaticalization on the frequency frequency distribution of Fijian in generation distributions in the following steps: 0 still reflects the original ZM values, the Simulation: Our program takes a given ZM parameter values after 6 generations of text of generation i, calculates a frequency grammaticalization have become much closer distribution for this generation, changes the to the values of the Hungarian UDHR: text along various operations given below, and gives the frequency distribution of the text for Fij (nG = 0): α = 1.21,β = 2.1,C = 812 a new generation i + 1 as output.
Recommended publications
  • Phonological Optimization and Syntactic Variation: the Case of Optional That*
    BLS 32, No 1 2006. DOI: http://dx.doi.org/10.3765/bls.v32i1.3453 (published by the Berkeley Linguistics Society and the Linguistic Society of America) Phonological Optimization and Syntactic Variation: The Case of * Optional That T. FLORIAN JAEGER Stanford University 1. Introduction The principle of Phonology-Free Syntax (see papers in Zwicky 1969) states that syntax should not be influenced by phonological constraints. This predicts that choice in syntactic variation should not be affected by phonological constraints, such as the avoidance of identical adjacent elements (Frisch et al. 2004; Leben 1973). Only a few studies have investigated the extent to which phonological encoding can influence syntactic encoding and they have come to conflicting results (see Bock and Eberhard 1993; Haskell and MacDonald 2003; MacDonald et al. 1993). I investigate the influence of phonological optimization at different levels of linguistic form on so called optional that, as in the following comple- 1 ment clauses (CC*s) and relative clauses (RC*s): (1a) You would think [CC* (that) there's no place for capital punishment in a ci- vilized western country]. (1b) I heard [CC* (that) that’s the second happiest day of your life]. (2a) I mean [NP any drugs [RC* (that) you want _ ]]. (2b) I mean [NP the way [RC* (that) it vibrates _ ]]. * I am grateful for the discussions that lead to this paper. I would like to thank D. Pesetsky, N. Richards, and A. Marantz for raising the question whether optional that is allomorphy rather than syntactic variation. I am particularly thankful to A. Antilla and L.
    [Show full text]
  • Phonological Evidence for Exemplar Storage of Mul Tiword Sequences
    SSLA. 24. 215-221 DOl 111.1111; S02722ti:llll21111211til PHONOLOGICAL EVIDENCE FOR EXEMPLAR STORAGE OF MULTIWORD SEQUENCES Joan Bybee University of New Mexico Phonological evidence supports the frequency-based model pro­ posed in the article by Nick Ellis. Phonological reduct1on occurs ear­ lier and to a greater extent in high-frequency words and phrases than in low-frequency ones. A model that accounts for this effect needs both an exemplar representation to show phonet1c variation and the ability to represent multiword combinations. The maintenance of alternations conditioned by word boundaries. such as French liai­ son, also provides evidence that multiword sequences are stored and can accrue representational strength. The reorganization of pho­ netic exemplars in favor of the more frequent types prov1des evidence for some abstraction in categories beyond the simple registration of tokens of experience. Ellis's review article paints a very different landscape for linguistic knowledge from that provided by structural theories of language. In this landscape, fre­ quency of use is not a constraint added to an otherwise structural or genera­ tive architecture (as in Hammond, 1999), nor is it just invoked when the facts do not yield entirely to structural analysis (see, e.g., Kaisse, 1985, p. 59). Rather, the picture that emerges from the array of studies that Ellis cites is one in which linguistic knowledge is based firmly on experience with language, and frequency of use is a foundational determinate of grammatical properties of language. The article demonstrates that there is a major convergence of re­ search from many different perspectives-corpus-based analysis, computational linguistics, discourse, cognitive and functional linguistics, and psycholinguis­ tics-that all point to a new theory of grammar with its attendant theory of language acquisition.
    [Show full text]
  • Joan Bybee (2001). Phonology and Language Use. (Cambridge Studies in Linguistics 94.) Cambridge: Cambridge University Press
    Phonology http://journals.cambridge.org/PHO Additional services for Phonology: Email alerts: Click here Subscriptions: Click here Commercial reprints: Click here Terms of use : Click here Joan Bybee (2001). Phonology and language use. (Cambridge Studies in Linguistics 94.) Cambridge: Cambridge University Press. Pp. xviii +238. Janet Pierrehumbert Phonology / Volume 19 / Issue 03 / December 2002, pp 459 - 463 DOI: 10.1017/S0952675703214445, Published online: 04 July 2003 Link to this article: http://journals.cambridge.org/abstract_S0952675703214445 How to cite this article: Janet Pierrehumbert (2002). Joan Bybee (2001). Phonology and language use. (Cambridge Studies in Linguistics 94.) Cambridge: Cambridge University Press. Pp. xviii+238.. Phonology, 19, pp 459-463 doi:10.1017/S0952675703214445 Request Permissions : Click here Downloaded from http://journals.cambridge.org/PHO, IP address: 129.105.215.146 on 09 Mar 2015 Phonology 19 (2002) 459–463. f 2002 Cambridge University Press DOI: 10.1017/S0952675703214445 Printed in the United Kingdom Reviews Joan Bybee (2001). Phonology and language use. (Cambridge Studies in Linguistics 94.) Cambridge: Cambridge University Press. Pp. xviii+238. Janet Pierrehumbert Northwestern University The functionalist viewpoint in linguistics can take different forms. A caricature of functionalist thinking is the notion that the structure of language is opti- mised, or nearly so, for its function as a means of human communication. This notion has met with widespread scepticism because of its lack of predictiveness in the face of typological variation. Either it leads to the prediction that all languages are en route to some single ‘Utopian’ (even if they have not quite achieved it) or it leads one to posit so many contradictory functional goods that the nature of possible languages is not effectively restricted.
    [Show full text]
  • Crosslinguistic Structural
    CROSSLINGUISTIC STRUCTURAL PRIMING AS A MECHANISM OF CONTACT-INDUCED LANGUAGE CHANGE: EVIDENCE FROM PAPIAMENTO-DUTCH BILINGUALS IN ARUBA AND THE NETHERLANDS Gerrit Jan Kootstra Hülya Şahin Centre for Language Studies, Radboud Centre for Language Studies, Radboud University Nijmegen University Nijmegen Studies on language contact suggest that cross-language interactions in individual language use may lead to contact-induced change at the community level. We propose that the phenomenon of crosslinguistic structural priming may well drive this process. We investigated this by focusing on dative sentence production by Papiamento speakers in Aruba and in the Netherlands. In experi - ment 1, Papiamento speakers in Aruba and in the Netherlands described dative events. The speak - ers in the Netherlands produced more Dutch-like structures than the speakers in Aruba, especially younger speakers. In experiment 2, speakers from the same populations heard a Dutch prime sen - tence before describing a dative event in Papiamento. Syntactic choices were influenced by the Dutch prime sentences, and, again, especially younger speakers in the Netherlands produced more Dutch-like dative structures. This combination of results suggests that Papiamento syntactic pref - erences in the Netherlands are changing as a function of contact with Dutch, and that crosslinguis - tic structural priming is a likely mechanism underlying this change.* Keywords : crosslinguistic priming, structural priming, contact-induced language change, bilin - gualism, dative alternation 1. Introduction. When multiple languages are spoken in the same situation and/or by the same people, this can lead to interaction between these languages, also called lan - guage contact. Language contact can be divided into ‘synchronic’ processes, like cross- language effects in bilingual language processing (e.g.
    [Show full text]
  • Alternatives to the Combinatorial Paradigm of Linguistic Theory Based on Domain General Principles of Human Cognition1
    Alternatives to the combinatorial paradigm of linguistic theory based on domain general principles of human cognition1 JOAN BYBEE AND JAMES L. MCCLELLAND Abstract It is argued that the principles needed to explain linguistic behavior are do- main-general and based on the impact that specific experiences have on the mental organization and representation of language. This organization must be sensitive to both specific information and generalized patterns. In addition, knowledge of language is highly sensitive to frequency of use: frequently-used linguistic sequences become more frequent, more accessible and better inte- grated. The evidence adduced is mainly from phonology and morphology and addresses the issue of gradience and specificity found in postulated units, cat- egories, and dichotomies such as regular and irregular, but the points apply to all levels of linguistic analysis including the syntactic, semantic, and discourse levels. Appropriate models for representing such phenomena are considered, including exemplar models and connectionist models, which are evolving to achieve a better fit with linguistic data. The major criticism of connectionist models often raised from within the combinatorial paradigm of much existing linguistic theory – that they do not capture ‘free combination’ to the extent that rule-based systems do, is regarded as a strength rather than a weakness. Re- cent connectionist models exhibit greater productivity and systematicity than earlier variants, but still show less uniformity of generalization than combina- torial models do. The remaining non-uniformity that the connectionist models show is appropriate, given that such non-uniformity is the rule in language structure and language behavior. 1. This article reflects initial stages of a dialogue between the authors, and the order of authorship is alphabetical.
    [Show full text]
  • Explanation in Typology
    Explanation in typology Diachronic sources, functional motivations and the nature of the evidence Edited by Karsten Schmidtke-Bode Natalia Levshina Susanne Maria Michaelis Ilja A. Seržant language Conceptual Foundations of science press Language Science 3 Conceptual Foundations of Language Science Series editors Mark Dingemanse, Max Planck Institute for Psycholinguistics N. J. Enfield, University of Sydney Editorial board Balthasar Bickel, University of Zürich, Claire Bowern, Yale University, Elizabeth Couper-Kuhlen, University of Helsinki, William Croft, University of New Mexico, Rose-Marie Déchaine, University of British Columbia, William A. Foley, University of Sydney , William F. Hanks, University of California at Berkeley, Paul Kockelman, Yale University, Keren Rice, University of Toronto, Sharon Rose, University of California at San Diego, Frederick J. Newmeyer, University of Washington, Wendy Sandler, University of Haifa, Dan Sperber Central European University No scientific work proceeds without conceptual foundations. In language science, our concepts about language determine our assumptions, direct our attention, and guide our hypotheses and our reason- ing. Only with clarity about conceptual foundations can we pose coherent research questions, design critical experiments, and collect crucial data. This series publishes short and accessible books that explore well-defined topics in the conceptual foundations of language science. The series provides a venue for conceptual arguments and explorations that do not require the traditional book-length treatment, yet that demand more space than a typical journal article allows. In this series: 1. Enfield, N. J. Natural causes of language. 2. Müller, Stefan. A lexicalist account of argument structure: Template-based phrasal LFG approaches and a lexical HPSG alternative 3. Schmidtke-Bode, Karsten, Natalia Levshina, Susanne Maria Michaelis & Ilja A.
    [Show full text]
  • Quantitative Determinants of Prefabs: a Corpus-Based, Experimental Study of Multiword Units in the Lexicon." (2013)
    University of New Mexico UNM Digital Repository Linguistics ETDs Electronic Theses and Dissertations 7-1-2013 Quantitative determinants of prefabs: A corpus- based, experimental study of multiword units in the lexicon Clayton Beckner Follow this and additional works at: https://digitalrepository.unm.edu/ling_etds Recommended Citation Beckner, Clayton. "Quantitative determinants of prefabs: A corpus-based, experimental study of multiword units in the lexicon." (2013). https://digitalrepository.unm.edu/ling_etds/3 This Dissertation is brought to you for free and open access by the Electronic Theses and Dissertations at UNM Digital Repository. It has been accepted for inclusion in Linguistics ETDs by an authorized administrator of UNM Digital Repository. For more information, please contact [email protected]. i Clayton Beckner Candidate Linguistics Department This dissertation is approved, and it is acceptable in quality and form for publication: Approved by the Dissertation Committee: Jill Morford, Chairperson Joan Bybee William Croft Andrew Wedel ii QUANTITATIVE DETERMINANTS OF PREFABS: A CORPUS-BASED, EXPERIMENTAL STUDY OF MULTIWORD UNITS IN THE LEXICON by CLAYTON BECKNER B.A., Physics, Bradley University, 1994 M.S., English, Illinois State University, 1999 M.A., Linguistics, University of New Mexico, 2005 DISSERTATION Submitted in Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy Linguistics The University of New Mexico Albuquerque, New Mexico July 2013 iii ACKNOWLEDGEMENTS This dissertation would not have been possible without the support of many individuals, both personally and academically. I thank, first of all, the many participants who generously volunteered their time for my studies, since I could not pursue empirical research without their help.
    [Show full text]
  • Language Change Joan Bybee Frontmatter More Information
    Cambridge University Press 978-1-107-02016-0 — Language Change Joan Bybee Frontmatter More Information Language Change How and why do languages change? This new introduction offers a guide to the types of change at all levels of linguistic structure, as well as the mechanisms behind each type. Based on data from a variety of methods and a huge array of language families, it examines general patterns of change, bringing together recent findings on sound change, analogical change, grammaticalization, the creation and change of constructions, as well as lexical change. Emphasizing crosslinguistic patterns and going well beyond traditional methods in historical linguistics, this book sees change as grounded in cognitive processes and usage factors that are rarely mentioned in other textbooks. Complete with questions for discussion, suggested readings, and a useful glossary of terms, this book helps students to gain a general understanding of language as an ever- changing system. joan bybee is Distinguished Professor Emerita of Linguistics at the University of New Mexico. © in this web service Cambridge University Press www.cambridge.org Cambridge University Press 978-1-107-02016-0 — Language Change Joan Bybee Frontmatter More Information CAMBRIDGE TEXTBOOKS IN LINGUISTICS General editors: p. austin, j. bresnan, b. comrie, s. crain, w. dressler, c. ewen, r. lass, d. lightfoot, k. rice, i. roberts, s. romaine, n.v. smith Language Change © in this web service Cambridge University Press www.cambridge.org Cambridge University Press 978-1-107-02016-0 — Language Change Joan Bybee Frontmatter More Information In this series: r. cann Formal Semantics j. laver Principles of Phonetics f.
    [Show full text]
  • Linguistic Typology 2019; 23(2): 263–302
    Linguistic Typology 2019; 23(2): 263–302 Joan Bybee and Shelece Easterday Consonant strengthening: A crosslinguistic survey and articulatory proposal https://doi.org/10.1515/lingty-2019-0015 Received April 16, 2018; accepted December 07, 2018 Abstract: Given the common intuition that consonant lenition occurs more often than fortition, we formulate this as a hypothesis, defining these sound change types in terms of decrease or increase in oral constriction. We then test the hypothesis on allophonic processes in a diverse sample of 81 languages. With the hypothesis confirmed, we examine the input and output of such sound changes in terms of manner and place of articulation and find that while decrease in oral constriction (weakening) affects most consonant types, increase in oral constriction (strengthening) is largely restricted to palatal and labial glides. We conclude that strengthening does not appear to be the simple inverse of weakening. In conclusion we suggest some possible avenues for explaining how glide strengthening may result from articulatory production pressures and speculate that strengthening and weakening can be encompassed under a single theory of sound change resulting from the automatization of production. Keywords: phonology, consonant strengthening, lenition and fortition, allopho- nic variation, sound change, automatization 1 Introduction It is generally agreed that lenition is much more common than fortition in sound changes as well as phonological processes across the languages of the world.1 This asymmetry suggests the need for special attention to fortition in order to determine if it is the simple inverse of lenition or a different phenomenon. The Author Contribution Statement: Joan Bybee and Shelece Easterday contributed equally to the data collection, analysis, and writing of this paper 1 See Section 3 for references and discussion.
    [Show full text]
  • Running Head: HOW EFFICIENCY SHAPES HUMAN LANGUAGE 1
    Running head: HOW EFFICIENCY SHAPES HUMAN LANGUAGE 1 How Efficiency Shapes Human Language Edward Gibson1,∗, Richard Futrell2, Steven T. Piantadosi3, Isabelle Dautriche4, Kyle Mahowald, Leon Bergen5, & Roger Levy1,∗ 1Massachusetts Institute of Technology, 2University of California, Irvine, 3University of California, Berkeley, 4University of Edinburgh, 5University of California, San Diego, ∗Corresponding authors: [email protected]; [email protected] Abstract Cognitive science applies diverse tools and perspectives to study human language. Recently, an exciting body of work has examined linguistic phenomena through the lens of efficiency in usage: what otherwise puzzling features of language find explanation in formal accounts of how language might be optimized for communication and learning? Here, we review studies that deploy formal tools from probability and information theory to understand how and why language works the way that it does, focusing on phenomena ranging from the lexicon through syntax. These studies show how a pervasive pressure for efficiency guides the forms of natural language and indicate that a rich future for language research lies in connecting linguistics to cognitive psychology and mathematical theories of communication and inference. Keywords: language evolution, communication, language efficiency, cross-linguistic universals, language learnability, language complexity HOW EFFICIENCY SHAPES HUMAN LANGUAGE 2 How Efficiency Shapes Human Language Why do languages look the way they do? Depending on how you count, there are 6000-8000 distinct languages on earth. The differences among languages seem pretty salient when you have to learn a new language: you have to learn a new set of sounds, a new set of words, and a new way of putting the words together; nevertheless, linguists have documented a wealth of strong recurring patterns across languages [1, 2].
    [Show full text]
  • Copyrighted Material
    Contents Notes on Contributors vii Acknowledgments xiii Introduction: Language Emergence 1 Brian MacWhinney Part I Basic Language Structures 33 1 The Emergence of Phonological Representation 35 Patricia Donegan 2 Capturing Gradience, Continuous Change, and Quasi-Regularity in Sound, Word, Phrase, and Meaning 53 James L. McClelland 3 The Emergence of Language Comprehension 81 Maryellen C. MacDonald 4 Anaphora and the Case for Emergentism 100 William O’Grady 5 Morphological Emergence 123 Péter Rácz, Janet B. Pierrehumbert, Jennifer B. Hay, and Viktória Papp 6 Metaphor and Emergentism 147 Zoltán Kövecses 7 Usage-Based LanguageCOPYRIGHTED Learning MATERIAL 163 Nick C. Ellis, Matthew Brook O’Donnell, and Ute Römer Part II Language Change and Typology 181 8 Emergence at the Cross-Linguistic Level: Attractor Dynamics in Language Change 183 Joan Bybee and Clay Beckner 9 The Diachronic Genesis of Synchronic Syntax 201 T. Givón vi Contents 10 Typological Variation and Efficient Processing 215 John A. Hawkins 11 Word Meanings across Languages Support Efficient Communication 237 Terry Regier, Charles Kemp, and Paul Kay Part III Interactional Structures 265 12 Linguistic Emergence on the Ground: A Variationist Paradigm 267 Shana Poplack and Rena Torres Cacoullos 13 The Emergence of Sociophonetic Structure 292 Paul Foulkes and Jennifer B. Hay 14 An Emergentist Approach to Grammar 314 Paul J. Hopper 15 Common Ground 328 Eve V. Clark 16 The Role of Culture in the Emergence of Language 354 Daniel L. Everett Part IV Language Learning 377 17 Learnability 379 Alexander Clark 18 Perceptual Development and Statistical Learning 396 Erik Thiessen and Lucy Erickson 19 Language Emergence in Development: A Computational Perspective 415 Stewart M.
    [Show full text]
  • First Language Acquisition
    University of Iceland School of Humanities Department of English First Language Acquisition Is Children’s knowledge of language innate? B. A. Essay Zulaia Johnston Da Cruz Kt.: 081279-2339 Supervisor: Þórhallur Eyþórsson May 2015 Abstract The aim of this essay is to examine the extent to which children’s language acquisition is innate. As such, this thesis highlights Noam Chomsky’s Innateness Hypothesis as the main theory underlying first language acquisition. Chomsky believes that children are born with an inherited capacity to learn languages. He points out that there is no way that children could learn language trough imitating adult speech because the speech they hear is not sufficient to aid children’s acquisition of language. He also claims that the only way to explain how children acquire the complex system of language is if they are born with an innate mechanism which aids the acquisition of language. A language acquisition device called Universal Grammar. Universal Grammar provides children with universal language principles and its grammatical structures. If Chomsky’s hypothesis is correct, then one can expect to find in human biology and development evidence that reflects specialization for language. Thus, this thesis discusses two ways in which humans are specialized for language. First, the thesis discusses the brain structure and how certain structures of children’s brains appear to be specialized for language. Secondly, we discuss the critical period for first language acquisition and its implications for children’s language acquisition after puberty. Furthermore, the present thesis assesses as well the controversy surrounding Chomsky’s hypothesis. Therefore, the criticism and theories of Jean Piaget, Michael Tomasello, Joan Bybee and Hilary Putnam are discussed.
    [Show full text]