The Minimalist Program in Syntax
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
On the Irrelevance of Transformational Grammar to Second Language Pedagogy
ON THE IRRELEVANCE OF TRANSFORMATIONAL GRAMMAR TO SECOND LANGUAGE PEDAGOGY John T. Lamendella The University of Michigan Many scholars view transformational grammar as an attempt to represent the structure of linguistic knowledge in the mind and seek to apply transformational descriptions of languages to the de- velopment of second language teaching materials. It will be claimed in this paper that it is a mistake to look to transformational gram- mar or any other theory of linguistic description to provide the theoretical basis for either second language pedagogy or a theory of language acquisition. One may well wish to describe the ab- stract or logical structure of a language by constructing a trans- formational grammar which generates the set of sentences identi- fied with that language. However, this attempt should not be con- fused with an attempt to understand the cognitive structures and processes involved in knowing or using a language. It is a cogni- tive theory of language within the field of psycholinguistics rather than a theory of linguistic description which should underlie lan- guage teaching materials. A great deal of effort has been expended in the attempt to demonstrate the potential contributions of the field of descriptive linguistics to the teaching of second languages and, since the theory of transformational grammar has become the dominant theory in the field of linguistics, it is not surprising that applied linguists have sought to apply transformational grammar to gain new in- sights into the teaching of second languages. It will be claimed in this paper that it is a mistake to look to transformational grammar or any other theory of linguistic description to provide the theoretical basis for either second language pedagogy or a theory of language acquisition. -
Scaffolding Learning of Language Structures with Visual‐Syntactic Text
UC Irvine UC Irvine Previously Published Works Title Scaffolding learning of language structures with visual-syntactic text formatting Permalink https://escholarship.org/uc/item/6235t25b Journal British Journal of Educational Technology, 50(4) ISSN 0007-1013 Authors Park, Y Xu, Y Collins, P et al. Publication Date 2018 DOI 10.1111/bjet.12689 Peer reviewed eScholarship.org Powered by the California Digital Library University of California British Journal of Educational Technology Vol 0 No 0 2018 1–17 doi:10.1111/bjet.12689 Scaffolding learning of language structures with visual-syntactic text formatting Youngmin Park, Ying Xu , Penelope Collins, George Farkas and Mark Warschauer Youngmin Park is a Lecturer at Pusan National University, Korea. She received her Ph.D. degree from the University of California, Irvine, specializing in Language, Literacy and Technology. Ying Xu is a Ph.D. student at the University of California, Irvine, with a specialization in specializing in Language, Literacy and Technology. Penelope Collins is an associate professor at the University of California, Irvine. Her research examines the development of language and literacy skills for children from linguistically diverse backgrounds. George Farkas is a professor at the University of California, Irvine. He has employed a range of statistical approaches and databases to examine the causes and consequences of reading achievement gap across varying age groups and educational settings. Mark Warschauer is a professor at the University of California, Irvine. He works on a range of research projects related to digital media in education. Address for correspondence: Ying Xu, University of California Irvine, 3200 Education Bldg, Irvine, CA 92697, USA. -
Word Order, Parameters and the Extended COMP
Alice Davison, University of Iowa [email protected] January 2006 Word order, parameters, and the Extended COMP projection Abstract The structure of finite CP shows some unexpected syntactic variation in the marking of finite subordinate clauses in the Indic languages, which otherwise are strongly head-final.. Languages with relative pronouns also have initial complementizers and conjunctions. Languages with final yes/no question markers allow final complementizers, either demonstratives or quotative participles. These properties define three classes, one with only final CP heads (Sinhala), one with only initial CP heads (Hindi, Panjabi, Kashmiri) and others with both possibilities. The lexical differences of final vs initial CP heads argue for expanding the CP projection into a number of specialized projections, whose heads are all final (Sinhala), all initial, or mixed. These projections explain the systematic variation in finite CPs in the Indic languages, with the exception of some additional restrictions and anomalies in the Eastern group. 1. Introduction In this paper, I examine two topics in the syntactic structure of clauses in the Indic languages. The first topic has to do with the embedding of finite clauses and especially about how embedded finite clauses are morphologically marked. The second topic focuses on patterns of linear order in languages, parameters of directionality in head position. The two topics intersect in the position of these markers of finite subordinate clauses in the Indic languages. These markers can be prefixes or suffixes, and I will propose that they are heads of functional projections, just as COMP is traditionally regarded as head of CP. The Indic languages are all fundamentally head-final languages; the lexically heads P, Adj, V and N are head-final in the surface structure, while only the functional head D is not. -
Serial Verb Constructions Revisited: a Case Study from Koro
Serial Verb Constructions Revisited: A Case Study from Koro By Jessica Cleary-Kemp A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Linguistics in the Graduate Division of the University of California, Berkeley Committee in charge: Associate Professor Lev D. Michael, Chair Assistant Professor Peter S. Jenks Professor William F. Hanks Summer 2015 © Copyright by Jessica Cleary-Kemp All Rights Reserved Abstract Serial Verb Constructions Revisited: A Case Study from Koro by Jessica Cleary-Kemp Doctor of Philosophy in Linguistics University of California, Berkeley Associate Professor Lev D. Michael, Chair In this dissertation a methodology for identifying and analyzing serial verb constructions (SVCs) is developed, and its application is exemplified through an analysis of SVCs in Koro, an Oceanic language of Papua New Guinea. SVCs involve two main verbs that form a single predicate and share at least one of their arguments. In addition, they have shared values for tense, aspect, and mood, and they denote a single event. The unique syntactic and semantic properties of SVCs present a number of theoretical challenges, and thus they have invited great interest from syntacticians and typologists alike. But characterizing the nature of SVCs and making generalizations about the typology of serializing languages has proven difficult. There is still debate about both the surface properties of SVCs and their underlying syntactic structure. The current work addresses some of these issues by approaching serialization from two angles: the typological and the language-specific. On the typological front, it refines the definition of ‘SVC’ and develops a principled set of cross-linguistically applicable diagnostics. -
Antisymmetry Kayne, Richard (1995)
CAS LX 523 Syntax II (1) A Spring 2001 March 13, 2001 qp Paul Hagstrom Week 7: Antisymmetry BE 33 Kayne, Richard (1995). The antisymmetry of syntax. Cambridge, MA: MIT Press. CDFG 1111 Koopman, Hilda (2000). The spec-head configuration. In Koopman, H., The syntax of cdef specifiers and heads. London: Routledge. (2) A node α ASYMMETRICALLY C-COMMANDS β if α c-commands β and β does not The basic proposals: c-command α. X-bar structures (universally) have a strict order: Spec-head-complement. There is no distinction between adjuncts and specifiers. • B asymmetrically c-commands F and G. There can be only one specifier. • E asymmetrically c-commands C and D. • No other non-terminal nodes asymmetrically c-command any others. But wait!—What about SOV languages? What about multiple adjunction? Answer: We’ve been analyzing these things wrong. (3) d(X) is the image of a non-terminal node X. Now, we have lots of work to do, because lots of previous analyses relied on d(X) is the set of terminal nodes dominated by node X. the availability of “head-final” structures, or multiple adjunction. • d(C) is {c}. Why make our lives so difficult? Wasn’t our old system good enough? • d(B) is {c, d}. Actually, no. • d(F) is {e}. A number of things had to be stipulated in X-bar theory (which we will review); • d(E) is {e, f}. they can all be made to follow from one general principle. • d(A) is {c, d, e, f}. The availability of a head-parameter actually fails to predict the kinds of languages that actually exist. -
Chapter 7 Linguistics As a Science of Structure Ryan M
Chapter 7 Linguistics as a science of structure Ryan M. Nefdt University of the Western Cape Generative linguistics has rapidly changed during the course of a relatively short period. This has caused many to question its scientific status as a realist scientific theory (Stokhof & van Lambalgen 2011; Lappin et al. 2000). In this chapter, I argue against this conclusion. Specifically, I claim that the mathematical foundations of the science present a different story below the surface. I agree with critics that due to the major shifts in theory over the past 80 years, linguistics is indeed opened up to the problem of pessimistic meta-induction or radical theory change. However, I further argue that a structural realist approach (Ladyman 1998; French 2006) can save the field from this problem and at the same time capture its structural nature. I discuss particular historical instances of theory change in generative grammar as evidence for this interpretation and finally attempt to extend it beyond the gener- ative tradition to encompass previous frameworks in linguistics. 1 Introduction The generativist revolution in linguistics started in the mid-1950s, inspired in large part by insights from mathematical logic and in particular proof theory. Since then, generative linguistics has become a dominant paradigm, with many connections to both the formal and natural sciences. At the centre of the newly established discipline was the syntactic or formal engine, the structures of which were revealed through modelling grammatical form. The generativist paradigm in linguistics initially relied heavily upon the proof-theoretic techniques intro- duced by Emil Post and other formal logicians to model the form language takes (Tomalin 2006; Pullum 2011; 2013).1 Yet despite these aforementioned formal be- ginnings, the generative theory of linguistics has changed its commitments quite 1Here my focus will largely be on the formal history of generative syntax but I will make some comments on other aspects of linguistics along the way. -
Syntactic Structures and Their Symbiotic Guests. Notes on Analepsis from the Perspective of On- Line Syntax*
Pragmatics 24:3.533-560 (2014) International Pragmatics Association DOI: 10.1075/prag.24.3.05aue SYNTACTIC STRUCTURES AND THEIR SYMBIOTIC GUESTS. NOTES ON ANALEPSIS FROM THE PERSPECTIVE OF ON- LINE SYNTAX* Peter Auer Abstract The empirical focus of this paper is on utterances that re-use syntactic structures from a preceding syntactic unit. Next utterances of this type are usually treated as (coordination) ellipsis. It is argued that from an on-line perspective on spoken syntax, they are better described as structural latency: A grammatical structure already established remains available and can therefore be made use of with one or more of its slots being filled by new material. A variety of cases of this particular kind of conversational symbiosis are discussed. It is argued that they should receive a common treatment. A number of features of the general host/guest relationship are discussed. Keywords: Analepsis; On-line syntax; Structural latency. "Ein Blick in die erste beste Erzählung und eine einfache Ueberlegung muss beweisen, dass jede frühere Aeusserung des Erzählenden die Exposition aller nachfolgenden Prädikate bildet." (Wegener 1885: 46)1 1. Introduction The notion of 'ellipsis' is often regarded with some skepticism by Interactional Linguists - the orientation to 'full' sentences is all too obvious (cf. Selting 1997), and the idea that speakers produce complete sentences just in order to delete some parts of them afterwards surely fails to account for the processual dynamics of sentence production and understanding (cf. Kindt 1985) in time. On the other hand, there can be little doubt that speakers often produce utterance units that could not be understood * I wish to thank Elizabeth Couper-Kuhlen and Susanne Günthner for their helpful comments on a previous version of this paper. -
LING83600: Context-Free Grammars
LING83600: Context-free grammars Kyle Gorman 1 Introduction Context-free grammars (or CFGs) are a formalization of what linguists call “phrase structure gram- mars”. The formalism is originally due to Chomsky (1956) though it has been independently discovered several times. The insight underlying CFGs is the notion of constituency. Most linguists believe that human languages cannot be even weakly described by CFGs (e.g., Schieber 1985). And in formal work, context-free grammars have long been superseded by “trans- formational” approaches which introduce movement operations (in some camps), or by “general- ized phrase structure” approaches which add more complex phrase-building operations (in other camps). However, unlike more expressive formalisms, there exist relatively-efficient cubic-time parsing algorithms for CFGs. Nearly all programming languages are described by—and parsed using—a CFG,1 and CFG grammars of human languages are widely used as a model of syntactic structure in natural language processing and understanding tasks. Bibliographic note This handout covers the formal definition of context-free grammars; the Cocke-Younger-Kasami (CYK) parsing algorithm will be covered in a separate assignment. The definitions here are loosely based on chapter 11 of Jurafsky & Martin, who in turn draw from Hopcroft and Ullman 1979. 2 Definitions 2.1 Context-free grammars A context-free grammar G is a four-tuple N; Σ; R; S such that: • N is a set of non-terminal symbols, corresponding to phrase markers in a syntactic tree. • Σ is a set of terminal symbols, corresponding to words (i.e., X0s) in a syntactic tree. • R is a set of production rules. -
Generative Linguistics and Neural Networks at 60: Foundation, Friction, and Fusion*
Generative linguistics and neural networks at 60: foundation, friction, and fusion* Joe Pater, University of Massachusetts Amherst October 3, 2018. Abstract. The birthdate of both generative linguistics and neural networks can be taken as 1957, the year of the publication of foundational work by both Noam Chomsky and Frank Rosenblatt. This paper traces the development of these two approaches to cognitive science, from their largely autonomous early development in their first thirty years, through their collision in the 1980s around the past tense debate (Rumelhart and McClelland 1986, Pinker and Prince 1988), and their integration in much subsequent work up to the present. Although this integration has produced a considerable body of results, the continued general gulf between these two lines of research is likely impeding progress in both: on learning in generative linguistics, and on the representation of language in neural modeling. The paper concludes with a brief argument that generative linguistics is unlikely to fulfill its promise of accounting for language learning if it continues to maintain its distance from neural and statistical approaches to learning. 1. Introduction At the beginning of 1957, two men nearing their 29th birthdays published work that laid the foundation for two radically different approaches to cognitive science. One of these men, Noam Chomsky, continues to contribute sixty years later to the field that he founded, generative linguistics. The book he published in 1957, Syntactic Structures, has been ranked as the most influential work in cognitive science from the 20th century.1 The other one, Frank Rosenblatt, had by the late 1960s largely moved on from his research on perceptrons – now called neural networks – and died tragically young in 1971. -
1 Minimalism Minimalism Is the Name of the Predominant Approach In
Minimalism Minimalism is the name of the predominant approach in generative linguistics today. It was first introduced by Chomsky in his work The Minimalist Program (1995) and has seen several developments and changes since. The central idea of minimalism is that a linguistic theory should contain as few non-derived assumptions as possible. Many notions that had been developed in earlier generative theory, specifically the Government & Binding theory (GB), have been abandoned, in an attempt to derive them from more basic concepts. In Chomsky's (1995) view, minimalism is an implementation of the more general Principles and Parameters model. According to this language model, the human language capacity consists of a set of universal principles combined with a set of parameters. The principles are thought to be innate, which means that every language adheres to them. The parameters can be thought of as switches that can be set in two positions. Learning the syntax of one's native language, according to Chomsky's view, is a matter of acquiring the correct parameter settings for the language. Chomsky describes syntax as a cognitive system that connects two other cognitive systems: the conceptual-intentional system and the articulatory-perceptual system. Because syntax is linked to these two systems, the syntactic model defines two interface levels, one for each of them: Phonological Form (PF) is the interface to the articulatory-perceptual system, and Logical Form (LF) is the interface to the conceptual-intentional system. The grammar model is built up as follows: a clause is derived by selecting a set of lexical items from the lexicon. -
Serial Verb Constructions: Argument Structural Uniformity and Event Structural Diversity
SERIAL VERB CONSTRUCTIONS: ARGUMENT STRUCTURAL UNIFORMITY AND EVENT STRUCTURAL DIVERSITY A DISSERTATION SUBMITTED TO THE DEPARTMENT OF LINGUISTICS AND THE COMMITTEE ON GRADUATE STUDIES OF STANFORD UNIVERSITY IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF DOCTOR OF PHILOSOPHY Melanie Owens November 2011 © 2011 by Melanie Rachel Owens. All Rights Reserved. Re-distributed by Stanford University under license with the author. This work is licensed under a Creative Commons Attribution- Noncommercial 3.0 United States License. http://creativecommons.org/licenses/by-nc/3.0/us/ This dissertation is online at: http://purl.stanford.edu/db406jt2949 ii I certify that I have read this dissertation and that, in my opinion, it is fully adequate in scope and quality as a dissertation for the degree of Doctor of Philosophy. Beth Levin, Primary Adviser I certify that I have read this dissertation and that, in my opinion, it is fully adequate in scope and quality as a dissertation for the degree of Doctor of Philosophy. Joan Bresnan I certify that I have read this dissertation and that, in my opinion, it is fully adequate in scope and quality as a dissertation for the degree of Doctor of Philosophy. Vera Gribanov Approved for the Stanford University Committee on Graduate Studies. Patricia J. Gumport, Vice Provost Graduate Education This signature page was generated electronically upon submission of this dissertation in electronic format. An original signed hard copy of the signature page is on file in University Archives. iii Abstract Serial Verb Constructions (SVCs) are constructions which contain two or more verbs yet behave in every grammatical respect as if they contain only one. -
Generative Linguistics Within the Cognitive Neuroscience of Language Alec Marantz
Generative linguistics within the cognitive neuroscience of language Alec Marantz Department of Linguistics and Philosophy, MIT Standard practice in linguistics often obscures the connection between theory and data, leading some to the conclusion that generative linguistics could not serve as the basis for a cognitive neuroscience of language. Here the foundations and methodology of generative grammar are clarified with the goal of explaining how generative theory already functions as a reasonable source of hypotheses about the representation and computation of language in the mind and brain. The claims of generative theory, as exemplified, e.g., within Chomsky’s (2000) Minimalist Program, are contrasted with those of theories endorsing parallel architectures with independent systems of generative phonology, syntax and semantics. The single generative engine within Minimalist approaches rejects dual routes to linguistic representations, including possible extra-syntactic strategies for semantic structure-building. Clarification of the implications of this property of generative theory undermines the foundations of an autonomous psycholinguistics, as established in the 1970’s, and brings linguistic theory back to the center of a unified cognitive neuroscience of language. 2 1. The place of linguistics* The first decade of the 21st century should be a golden era for the cognitive neuroscience of language. Fifty years of contemporary linguistic analysis of language can be coupled with a wide range of brain imaging and brain monitoring machines to test hypotheses and refine theory and understanding. However, there is still a gulf between mainstream linguistics within the generative linguistic tradition and most of those engaged in experimental cognitive neuroscience research. Some have argued that the fault here lies with the linguists, whose generative theories are based in principle on separating the study of linguistic representations from research on the acquisition and use of language in the minds and brains of speakers.