Relating Structure and Time in Linguistics and Psycholinguistics Colin Phillips and Matthew Wagers

Total Page:16

File Type:pdf, Size:1020Kb

Relating Structure and Time in Linguistics and Psycholinguistics Colin Phillips and Matthew Wagers 45-Gaskell-Chap45 3/10/07 8:15 PM Page 739 CHAPTER 45 Relating structure and time in linguistics and psycholinguistics Colin Phillips and Matthew Wagers 45.1 Linguistics and questions, and vice versa. This is an area where a rich linguistic literature and a sizeable body of psycholinguistics psycholinguistic research address closely related The field of psycholinguistics advertises its men- phenomena. The most widely discussed form of talistic commitments in its name. The field of unbounded dependency occurs when a noun linguistics does not. Psycholinguistic research fre- phrase (NP) such as which voters in (1) appears in quently involves ingenious experimental designs, a position that is structurally distant from the fancy lab equipment such as eye-trackers or elec- verb that it is an argument of (e.g. bribe in (1)). troencephalograms (EEGs), large groups of exper- Following standard practice, we mark the canon- imental subjects, and detailed statistical analyses. ical position of the direct object of bribe with an Linguistic research typically requires no special- underline or gap in (1), but it is a matter of great ized equipment, no statistical analyses, and controversy whether such gap positions are a part somewhere between zero and a handful of coop- of the mental representation of sentences like (1). erative informants. Psycholinguistic research is After reviewing some of the competing linguistic most commonly conducted in a Department of analyses of unbounded dependency construc- Psychology. Linguistic research is not. Some of tions in section 45.2, we discuss in section 45.3 these differences may contribute to the wide- the contributions of psycholinguistics to the ques- spread perception, well-represented among lin- tion of how these dependencies are represented. guists and psychologists alike, that the concerns (1) Which voters did the prosecutor suspect of psycholinguistics are somehow more psycho- that the candidate wanted his operatives to logical than those of linguistics, and that psy- bribe ___ before the election? cholinguistics can be looked to for psychological validation of the constructs proposed by linguists. The status of constraints on long-distance Although this view of the relation between the dependencies, such as the ban on dependencies two fields gives the impression of a neat division that span relative clause boundaries, as illustrated of labor, we find it misleading, and suspect that in (2), is a major topic of linguistic research. it may have led to unrealistic expectations, and In section 45.4 we discuss the effects of such consequently to disappointments and mutual constraints on language processing and their frustration. implications for the relation between linguistic In this commentary we focus on issues in the and psycholinguistic models. representation of unbounded syntactic depend- (2) *Which voters did the prosecutor ask [NP encies, as a case study of what psycholinguistic the operative [RC that bribed ___]] to testify methods can and cannot tell us about linguistic against his boss? 45-Gaskell-Chap45 3/10/07 8:15 PM Page 740 740 · CHAPTER 45 Relating Structure and Time in Linguistics and Psycholinguistics As far as we can tell, there is no principled dif- tasks such as speaking and understanding. In ference between the psychological relevance of contrast, questions about realtime processes have psycholinguistic and linguistic research. Most been central in adult psycholinguistics, whereas modern linguists have serious mentalistic com- less attention has been given to the question of mitments (and those that do not are of little con- why certain expressions are possible and others cern to us here).1 The data that linguists and are impossible. This focus of adult psycholin- psycholinguists collect and the theories that they guistics upon mechanisms that are closely tied develop based on those data are all “psychologi- to speaking and understanding is sometimes jus- cal,” in the sense that they aim to explain some tified by the notion that these are more directly aspect of human cognitive abilities. There are cer- related to common behaviors, or by the assump- tainly differences in the issues and methods that tion that the goal of psycholinguistics is to pro- the two fields tend to pay closest attention to; but vide “processing models.” However, a look at we are unaware of reasons to think that either psycholinguistic work with children casts doubt discipline has more direct access to psychological upon this rationale. Developmental psycholin- evidence, and we would include in this those guistics has devoted much attention to the ques- strains of psycholinguistics that draw on cogni- tion of what children can and cannot represent tive neuroscience methods, as we do in our own at different ages. Studies of childrens realtime work. Since this is not a standard position, we will processing and detailed studies of learning have briefly attempt to substantiate this claim. become more prominent only recently. Much There is broad agreement that mastery of a lan- work in developmental psycholinguistics asks the guage involves at least an (unconscious) under- same questions about children that linguists ask standing of the range of possible expressions that about adults. Therefore, the question of what a can be represented in the language, and also an speaker can and cannot represent, and the ques- ability to identify those expressions that cannot be tion of how the representations are constructed represented in the language. It is also agreed that in time, are presumably both psychologically since the expressive power of human languages is respectable concerns. We suspect that the discipli- too large for any individual to simply memorize nary divisions have more to do with the method- the possible representations of his language, a ological biases of the respective fields. speaker must have the ability to generate, recog- There are obvious differences between the data nize, and interpret novel expressions of the lan- collection methods most commonly used in lin- guage relatively quickly in order to speak and guistics and psycholinguistics. The primary data understand. Since humans are able to learn any of theoretical linguistics comes from native speak- language for which they receive adequate expo- ers’ intuitive judgements about the acceptability sure from an early age, it is also agreed that the of sentences or the availability of specific inter- ability to learn language is a key component of the pretations. Such data are relatively easy to come human capacity for language. Understanding by, making it possible to establish a large number each of these abilities is an important part of the of facts about many different languages in rela- task of explaining how human language works, tively little time. Psycholinguistic data, on the and it is perhaps an accident of history or method- other hand, typically require a good deal more ology that different sub-fields have emerged that effort. In order to establish reliable generalizations focus on each of these problems. about reaction times, focal brain activity, or any of Linguistic theory typically focuses on charac- a number of other common dependent measures, terizing the representations that a speaker of a one needs to use specialized equipment, test large language can entertain, often allied with the ques- numbers of experimental items on large num- tion of what is a “possible human language.” bers of participants, devise ingenious ways to hide Theoretical linguists have generally paid less atten- one’s goals from the participants, and use com- tion to questions about how these representa- plex statistical analyses to interpret the results. tions might be retrieved or constructed in realtime It can take a lot of work to establish just one fact, and it can be difficult to conduct experiments on a number of different languages. 1 We should emphasize that we are concerned in this com- The different data collection practices of lin- mentary with what we take to be the “best practices” in either guistics and psycholinguistics have a clear impact field. It is not difficult to find instances of careless misrep- upon the fields. First, they affect the empirical resentation of linguistic data, uninterpretable experimental designs, unwarranted inferences from brain activation scope of the fields. Thanks to its low-tech meth- patterns, etc., of which neither field would be proud. Our ods, linguistics has amassed a large body of find- interest here is more in what can be learned from carefully ings from a very diverse set of languages, including conducted work in either field. languages for which only a small number of 45-Gaskell-Chap45 3/10/07 8:15 PM Page 741 Linguistics and psycholinguistics · 741 speakers remain. In contrast, most psycholin- among half a dozen friends will do (e.g. two-tailed guistic research has been confined to a handful sign test). One can often do without the half dozen of closely related western European languages. friends, too, if one is sufficiently confident about Second, differences in data collection methods the judgement. There are, of course, examples of affect what linguists and psycholinguists spend errors and disagreements, and notorious cases their time on, shape what is valued in the two of judgements that are subtle at best, but these fields, and also affect the safeguards that the two are the exception rather than the rule. In our fields place on data reliability. In psycholinguis- own work we often run large acceptability rating tics data collection is sufficiently difficult that studies as controls for our on-line studies. These great value is placed on elegant experimental studies cost little effort, since the materials are designs that make it possible to establish a single independently needed for the on-line studies, fact, and many procedural safeguards are put in but the results are almost never surprising, and place in order to ensure that results are reliable. are generally so robust statistically as to indicate In linguistics, on the other hand, data collection that we tested more subjects than needed.
Recommended publications
  • The Power of Locality Domains in Phonology∗
    The Power of Locality Domains in Phonology∗ Thomas Graf Stony Brook University Abstract Domains play an integral role in linguistic theories. This paper com- bines locality domains with current work on the computational complex- ity of phonology. The first result is that if a specific formalism — Strictly Piecewise (SP) grammars — is supplemented with a mechanism to enforce first-order definable domain restrictions, its power increases so much that it subsumes almost the full hierarchy of subregular languages. However, if domain restrictions are based on linguistically natural intervals, one in- stead obtains an empirically more adequate model. On the on hand, this model subsumes only those subregular classes that have been argued to be relevant for phonotactic generalizations. On the other hand, it excludes unnatural generalizations that involve counting or elaborate conditionals. It is also shown that SP grammars with interval-based domains are theo- retically learnable unlike SP grammars with arbitrary, first-order domains. Keywords: locality, phonological domains, first-order logic, tiers, sub- regular hierarchy, strictly piecewise, learnability 1 Introduction It has been known for a long time that the phonological systems of natural lan- guages are finite-state in nature and thus generate regular languages (Johnson 1972; Kaplan and Kay 1994). However, this characterization is insufficent in the sense that phonological dependencies instantiate much simpler patterns — the ∗I am greatly indebted to two anonymous reviewers and the editors of this issue. Their de- tailed feedback led to major changes that, hopefully, have resulted in a much more accessible and phonologically grounded paper. This paper has also profited tremendously from regular discussions of subregular phonology with Alëna Aksënova, Hyunah Baek, Aniello De Santo, and Chikako Takahashi.
    [Show full text]
  • Scaffolding Learning of Language Structures with Visual‐Syntactic Text
    UC Irvine UC Irvine Previously Published Works Title Scaffolding learning of language structures with visual-syntactic text formatting Permalink https://escholarship.org/uc/item/6235t25b Journal British Journal of Educational Technology, 50(4) ISSN 0007-1013 Authors Park, Y Xu, Y Collins, P et al. Publication Date 2018 DOI 10.1111/bjet.12689 Peer reviewed eScholarship.org Powered by the California Digital Library University of California British Journal of Educational Technology Vol 0 No 0 2018 1–17 doi:10.1111/bjet.12689 Scaffolding learning of language structures with visual-syntactic text formatting Youngmin Park, Ying Xu , Penelope Collins, George Farkas and Mark Warschauer Youngmin Park is a Lecturer at Pusan National University, Korea. She received her Ph.D. degree from the University of California, Irvine, specializing in Language, Literacy and Technology. Ying Xu is a Ph.D. student at the University of California, Irvine, with a specialization in specializing in Language, Literacy and Technology. Penelope Collins is an associate professor at the University of California, Irvine. Her research examines the development of language and literacy skills for children from linguistically diverse backgrounds. George Farkas is a professor at the University of California, Irvine. He has employed a range of statistical approaches and databases to examine the causes and consequences of reading achievement gap across varying age groups and educational settings. Mark Warschauer is a professor at the University of California, Irvine. He works on a range of research projects related to digital media in education. Address for correspondence: Ying Xu, University of California Irvine, 3200 Education Bldg, Irvine, CA 92697, USA.
    [Show full text]
  • A Computational Model of Cognitive Constraints in Syntactic Locality
    A COMPUTATIONAL MODEL OF COGNITIVE CONSTRAINTS IN SYNTACTIC LOCALITY A Dissertation Presented to the Faculty of the Graduate School of Cornell University in Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy by Marisa Ferrara Boston January 2012 c 2012 Marisa Ferrara Boston ALL RIGHTS RESERVED A COMPUTATIONAL MODEL OF COGNITIVE CONSTRAINTS IN SYNTACTIC LOCALITY Marisa Ferrara Boston, Ph.D. Cornell University 2012 This dissertation is broadly concerned with the question: how do human cognitive limitations influence difficult sentences? The focus is a class of grammatical restrictions, locality con- straints. The majority of relations between words are local; the relations between question words and their governors are not. Locality constraints restrict the formation of these non-local dependencies. Though necessary, the origin, operation, and scope of locality constraints is a controversial topic in the literature. The dissertation describes the implementation of a computational model that clarifies these issues. The model tests, against behavioral data, a series of cognitive constraints argued to account for locality. The result is an explanatory model predictive of a variety of cross-linguistic locality data. The model distinguishes those cognitive limitations that affect locality processing, and addresses the competence-performance debate by determining how and when cognitive constraints explain human behavior. The results provide insight into the nature of locality constraints, and promote language models sensitive to human cognitive limitations. BIOGRAPHICAL SKETCH Marisa Ferrara Boston (nee´ Marisa Angela Ferrara) was born January 26, 1981 in Painesville, Ohio. She received a Bachelor of Arts degree in Individualized Studies from Eastern Michigan University in 2001, and a Master of Arts degree in English Linguistics from Eastern Michigan University in 2005.
    [Show full text]
  • Native American and Indigenous Philosophy
    NEWSLETTER | The American Philosophical Association Native American and Indigenous Philosophy SPRING 2020 VOLUME 19 | NUMBER 2 FROM THE MANAGING EDITOR Agnes B. Curry SUBMISSION GUIDELINES AND INFORMATION ARTICLE Andrea Sullivan-Clarke Empowering Relations: An Indigenous Understanding of Allyship BOOK REVIEWS Brian Burkhart: Indigenizing Philosophy through the Land: A Trickster Methodology for Decolonizing Environmental Ethics and Indigenous Futures Reviewed by Joseph Len Miller Indigenous Philosophy, Locality, and Dance: A Joint Review of Shay Welch, The Phenomenology of a Performance Knowledge System: Dancing with Native American Epistemology, and Brian Burkhart, Indigenizing Philosophy through the Land: A Trickster Methodology for Decolonizing Environmental Ethics and Indigenous Futures Reviewed by Dennis H. McPherson and J. Douglas Rabb Shay Welch: The Phenomenology of a Performance Knowledge System: Dancing with Native American Epistemology Reviewed by Lorraine Mayer VOLUME 19 | NUMBER 2 SPRING 2020 © 2020 BY THE AMERICAN PHILOSOPHICAL ASSOCIATION ISSN 2155-9708 APA NEWSLETTER ON Native American and Indigenous Philosophy AGNES B. CURRY, EDITOR VOLUME 19 | NUMBER 2 | SPRING 2020 FROM THE MANAGING EDITOR SUBMISSION GUIDELINES AND Agnes B. Curry INFORMATION UNIVERSITY OF SAINT JOSEPH, CONNECTICUT We invite you to submit your work for consideration for In this newsletter we offer first an article by Andrea Sullivan- publication in the Newsletter on Native American and Clarke of Windsor University in Canada. In “Empowering Indigenous Philosophy. We welcome comments and Relations” she critically considers recent discussions responses to work published in this or past issues. We also of allyship and the limitations inherent in them, so as to welcome work that speaks to philosophical, professional build a less ethically naïve concept for allying work with and community concerns regarding Native American and indigenous peoples and communities.
    [Show full text]
  • Locality Constraints It Is Instructive to Compare Overt Movement Like Wh-Movement and QR, Which Is a Type of Covert Movement
    PLIN3004/PLING218 Advanced Semantic Theory Lecture Notes: Week 10 Quantifiers sometimes take scope in places different from where they appear on the surface. One such case is the inverse scope reading of (1), which we discussed in Lecture 9. (1) Someone likes everyone. This shows that the scope of a quantifier is not completely determined by the surface form of the sentence. However, at the same time, it is known that quantifier scope is quite constrained in natural language. In order to understand such constraints, we adopt here the hypothesis that quantifier scope is determined by the covert movement operation Quantifier Raising (QR), and constraints on quantifier scope are constraints on QR. As we will see, the discussion is largely open-ended, but it is clear that there are some peculiar properties of quantifier scope in natural languages that we need to explain. 1 Locality Constraints It is instructive to compare overt movement like wh-movement and QR, which is a type of covert movement. As you have learned in syntax modules, wh-movement is known to be subject to a number of locality constraints known as island constraints. Here are two examples of island constraints: (2) Complex NP Island *Which novel did somebody meet a man that has read t? (3) Coordinate Structure Island *Which novel did somebody [read t and watched TV] before going to bed? Interestingly, quantifier scope seems to be subject to these constraints as well. For example, every novel cannot take scope over somebody in the following examples. (4) Complex NP Island Somebody met a man that has read every novel.
    [Show full text]
  • The Articulatory Basis of Locality in Phonology
    The Articulatory Basis of Locality in Phonology by Adamantios I. Gafos Notes: 1. This dissertation has been published by Garland Publishing, Inc., New York, 1999. 2. Copyright © 1999 Adamantios I. Gafos. ISBN 0-8153-3286-6 3. How to navigate through this document. The entire dissertation is comprised of 8 separate pdf files broken into chapters, bibliography and index for the benefit of users who have a slow internet connection. Hyperlinks to each chapter (Chapter 1, 2, 3, etc.), bibliography and index are found in the Contents (Table of Contents) page of this file. The hyperlinks are high-lighted with either a magenta or blue colored text. A hyperlink ( ) located on the upper right- hand corner of Page 1 of each chapter links that chapter back to this page. Dedication to my parents, Ioannis and Ioanna and to my sisters, Anthippi and Ioulia v vi Contents Preface ....................................................... xi Acknowledgments ............................................. xiii Illustrations ...................................................xv Abbreviations ................................................xvii Chapter 1: Introduction 1. Central Thesis ................................................3 2. Theoretical Background ........................................10 2.1 Gestures in Articulatory Phonology .....................11 2.2 Specific Assumptions ................................16 3. Organization of the Dissertation ................................ 21 Notes ........................................................24 Chapter
    [Show full text]
  • Chapter 7 Linguistics As a Science of Structure Ryan M
    Chapter 7 Linguistics as a science of structure Ryan M. Nefdt University of the Western Cape Generative linguistics has rapidly changed during the course of a relatively short period. This has caused many to question its scientific status as a realist scientific theory (Stokhof & van Lambalgen 2011; Lappin et al. 2000). In this chapter, I argue against this conclusion. Specifically, I claim that the mathematical foundations of the science present a different story below the surface. I agree with critics that due to the major shifts in theory over the past 80 years, linguistics is indeed opened up to the problem of pessimistic meta-induction or radical theory change. However, I further argue that a structural realist approach (Ladyman 1998; French 2006) can save the field from this problem and at the same time capture its structural nature. I discuss particular historical instances of theory change in generative grammar as evidence for this interpretation and finally attempt to extend it beyond the gener- ative tradition to encompass previous frameworks in linguistics. 1 Introduction The generativist revolution in linguistics started in the mid-1950s, inspired in large part by insights from mathematical logic and in particular proof theory. Since then, generative linguistics has become a dominant paradigm, with many connections to both the formal and natural sciences. At the centre of the newly established discipline was the syntactic or formal engine, the structures of which were revealed through modelling grammatical form. The generativist paradigm in linguistics initially relied heavily upon the proof-theoretic techniques intro- duced by Emil Post and other formal logicians to model the form language takes (Tomalin 2006; Pullum 2011; 2013).1 Yet despite these aforementioned formal be- ginnings, the generative theory of linguistics has changed its commitments quite 1Here my focus will largely be on the formal history of generative syntax but I will make some comments on other aspects of linguistics along the way.
    [Show full text]
  • Local Domains and Non-Local Dependencies in a Lexicalised Tree Adjoining Grammar
    philosophies Article Lexicalised Locality: Local Domains and Non-Local Dependencies in a Lexicalised Tree Adjoining Grammar Diego Gabriel Krivochen 1,* and Andrea Padovan 2 1 Dipartimento di Culture e Civiltà, Università di Verona, 37129 Verona, Italy 2 Dipartimento di Lingue e Letterature Straniere, Università di Verona, 37129 Verona, Italy; [email protected] * Correspondence: [email protected] Abstract: Contemporary generative grammar assumes that syntactic structure is best described in terms of sets, and that locality conditions, as well as cross-linguistic variation, is determined at the level of designated functional heads. Syntactic operations (merge, MERGE, etc.) build a structure by deriving sets from lexical atoms and recursively (and monotonically) yielding sets of sets. Additional restrictions over the format of structural descriptions limit the number of elements involved in each operation to two at each derivational step, a head and a non-head. In this paper, we will explore an alternative direction for minimalist inquiry based on previous work, e.g., Frank (2002, 2006), albeit under novel assumptions. We propose a view of syntactic structure as a specification of relations in graphs, which correspond to the extended projection of lexical heads; these are elementary trees in Tree Adjoining Grammars. We present empirical motivation for a lexicalised approach to structure building, where the units of the grammar are elementary trees. Our proposal will be based on cross-linguistic evidence; we will consider the structure of elementary trees in Spanish, English and German. We will also explore the consequences of assuming that nodes in elementary trees are Citation: Krivochen, D.G.; Padovan, addresses for purposes of tree composition operations, substitution and adjunction.
    [Show full text]
  • Information-Theoretic Locality Properties of Natural Language
    Information-theoretic locality properties of natural language Richard Futrell Department of Language Science University of California, Irvine [email protected] Abstract I present theoretical arguments and new empirical evidence for an information-theoretic principle of word order: information locality, the idea that words that strongly predict each other should be close to each other in linear order. I show that information locality can be derived under the assumption that nat- ural language is a code that enables efficient communication while minimizing information-processing costs involved in online language comprehension, using recent psycholinguistic theories to characterize those processing costs information-theoretically. I argue that information locality subsumes and extends the previously-proposed principle of dependency length minimization (DLM), which has shown great explanatory power for predicting word order in many languages. Finally, I show corpus evidence that information locality has improved explanatory power over DLM in two domains: in predicting which de- pendencies will have shorter and longer lengths across 50 languages, and in predicting the preferred order of adjectives in English. 1 Introduction The field of functional linguistics has long argued that the distinctive properties of natural language are best ex- plained in terms of what makes for an efficient communication system under the cognitive constraints particular to human beings. The idea is that the properties of language are determined by the pressure to enable efficient communication while minimizing the information-processing effort required for language production and compre- hension by humans. Within that field, a particularly promising concept is the principle of dependency length minimization (DLM): the idea that words linked in syntactic dependencies are under a pressure to be close in linear order.
    [Show full text]
  • Syntactic Structures and Their Symbiotic Guests. Notes on Analepsis from the Perspective of On- Line Syntax*
    Pragmatics 24:3.533-560 (2014) International Pragmatics Association DOI: 10.1075/prag.24.3.05aue SYNTACTIC STRUCTURES AND THEIR SYMBIOTIC GUESTS. NOTES ON ANALEPSIS FROM THE PERSPECTIVE OF ON- LINE SYNTAX* Peter Auer Abstract The empirical focus of this paper is on utterances that re-use syntactic structures from a preceding syntactic unit. Next utterances of this type are usually treated as (coordination) ellipsis. It is argued that from an on-line perspective on spoken syntax, they are better described as structural latency: A grammatical structure already established remains available and can therefore be made use of with one or more of its slots being filled by new material. A variety of cases of this particular kind of conversational symbiosis are discussed. It is argued that they should receive a common treatment. A number of features of the general host/guest relationship are discussed. Keywords: Analepsis; On-line syntax; Structural latency. "Ein Blick in die erste beste Erzählung und eine einfache Ueberlegung muss beweisen, dass jede frühere Aeusserung des Erzählenden die Exposition aller nachfolgenden Prädikate bildet." (Wegener 1885: 46)1 1. Introduction The notion of 'ellipsis' is often regarded with some skepticism by Interactional Linguists - the orientation to 'full' sentences is all too obvious (cf. Selting 1997), and the idea that speakers produce complete sentences just in order to delete some parts of them afterwards surely fails to account for the processual dynamics of sentence production and understanding (cf. Kindt 1985) in time. On the other hand, there can be little doubt that speakers often produce utterance units that could not be understood * I wish to thank Elizabeth Couper-Kuhlen and Susanne Günthner for their helpful comments on a previous version of this paper.
    [Show full text]
  • LING83600: Context-Free Grammars
    LING83600: Context-free grammars Kyle Gorman 1 Introduction Context-free grammars (or CFGs) are a formalization of what linguists call “phrase structure gram- mars”. The formalism is originally due to Chomsky (1956) though it has been independently discovered several times. The insight underlying CFGs is the notion of constituency. Most linguists believe that human languages cannot be even weakly described by CFGs (e.g., Schieber 1985). And in formal work, context-free grammars have long been superseded by “trans- formational” approaches which introduce movement operations (in some camps), or by “general- ized phrase structure” approaches which add more complex phrase-building operations (in other camps). However, unlike more expressive formalisms, there exist relatively-efficient cubic-time parsing algorithms for CFGs. Nearly all programming languages are described by—and parsed using—a CFG,1 and CFG grammars of human languages are widely used as a model of syntactic structure in natural language processing and understanding tasks. Bibliographic note This handout covers the formal definition of context-free grammars; the Cocke-Younger-Kasami (CYK) parsing algorithm will be covered in a separate assignment. The definitions here are loosely based on chapter 11 of Jurafsky & Martin, who in turn draw from Hopcroft and Ullman 1979. 2 Definitions 2.1 Context-free grammars A context-free grammar G is a four-tuple N; Σ; R; S such that: • N is a set of non-terminal symbols, corresponding to phrase markers in a syntactic tree. • Σ is a set of terminal symbols, corresponding to words (i.e., X0s) in a syntactic tree. • R is a set of production rules.
    [Show full text]
  • Generative Linguistics and Neural Networks at 60: Foundation, Friction, and Fusion*
    Generative linguistics and neural networks at 60: foundation, friction, and fusion* Joe Pater, University of Massachusetts Amherst October 3, 2018. Abstract. The birthdate of both generative linguistics and neural networks can be taken as 1957, the year of the publication of foundational work by both Noam Chomsky and Frank Rosenblatt. This paper traces the development of these two approaches to cognitive science, from their largely autonomous early development in their first thirty years, through their collision in the 1980s around the past tense debate (Rumelhart and McClelland 1986, Pinker and Prince 1988), and their integration in much subsequent work up to the present. Although this integration has produced a considerable body of results, the continued general gulf between these two lines of research is likely impeding progress in both: on learning in generative linguistics, and on the representation of language in neural modeling. The paper concludes with a brief argument that generative linguistics is unlikely to fulfill its promise of accounting for language learning if it continues to maintain its distance from neural and statistical approaches to learning. 1. Introduction At the beginning of 1957, two men nearing their 29th birthdays published work that laid the foundation for two radically different approaches to cognitive science. One of these men, Noam Chomsky, continues to contribute sixty years later to the field that he founded, generative linguistics. The book he published in 1957, Syntactic Structures, has been ranked as the most influential work in cognitive science from the 20th century.1 The other one, Frank Rosenblatt, had by the late 1960s largely moved on from his research on perceptrons – now called neural networks – and died tragically young in 1971.
    [Show full text]