The Minimalist Program (MP) in a Page: Review

Total Page:16

File Type:pdf, Size:1020Kb

The Minimalist Program (MP) in a Page: Review The Minimalist Program (MP) in a page: review Features drive syntax; they can be binary or multi-valued. Feature specifications are a huge part of syntax (and hence of our course). The MP focuses largely on uninterpretable features, though phi- features (which are largely interpretable), are important too (e.g. for agreement). Lexical categories, case, inflectional morphology, and subcategorization information all relate to features. The X’ schema is largely followed, but not religiously (cf. bare phrase structure, which we’ll study later). First and Second Merge are the cornerstone of structure. MP tries to describe why, not just how, these Merge operators take place. Binary branching is preferred by most. We will do a little semantics, too, and so predicates (and their meaning) are salient for us. Predicates have arity (adicity) that reflects their arguments and adjuncts. Logic and linguistics both use predicate expressions extensively. Arguments receive theta roles, another area of interaction between syntax and semantics. UTAH is important. The MP calls the featural requirements for subcategorization “c(ategorial)-selection”, and semantic restrictions “s(emantic)-selection”. Uninterpretable c-selection features must be satisfied (via checking, usually under sisterhood), and once checked they can delete. The feature Full Interpretation requires that all uninterpretable features must be deleted before the interface with semantics. (Remember, semantics is interpreted from prior syntax). The Extension Condition says you can only build a tree up from its root. If a feature can’t be checked where it is, it may percolate up to its mother node. Adjunction is allowed to happen at all phrasal levels. C-command can account for many syntactic phenomena crosslinguistically: binding pronouns, defining negative polarity environments, feature checking for agreement, etc. Ditransitives (including double object constructions) are problematic for binary X’. So are causatives, light verbs, etc. To accommodate them a new layer is added to the VP: v and its projections, aka the VP- shell). Intransitives can be split into unergatives and unaccusatives. There’s a lot of crosslinguistic evidence for this (e.g. ne-cliticization in Italian, French aux verb selection, etc.). Functional projections are X’ structures for function words (e.g. aux verbs, determiners) or non-lexical properties. X’ projections can be described hierarchically. Adger’s clausal hierarchy of projections for English is: (C) >I > (Neg) > (Perf) > (Prog) > (Pass) > v > V. Features can be either strong or weak. Strong features trigger movement. For example, modal verbs in I move to C for a yes-no question (YNQ) with do-support when necessary; the topmost aux verb moves to I, and some verbs (e.g. copulars) move to T. Other languages behave differently if relevant features aren’t strong. Agreement is done via checking and valuing features via c-command. Movements can be categorized several ways; two that are salient for us is what moves (head vs. phrasal) and where things move to (A vs. A-bar). Chains are sometimes relevant. So are restrictions on how far something can move (at once), e.g. subjacency. Derivations that aren’t successful before Spellout (the interface between syntax and PF and LF) crash. Subjects start in spec-v and then move to spec-IP (in languages with strong EPP). So does the unaccusative argument, the passive internal argument, NP is better expressed as DP, including a n shell. Many associated features include umass, uproper, gen, etc. Quantifier float, theta-assigning nominals, and genitive/possessive constructions are of interest. Plus classifiers in other languages. The C complex is for complementizers, and the uC and uclause-type features help determine clausal structure and function. Untensed complements are interesting from a syntactic and semantic point of view. PRO is introduced when a theta role has no constituent to absorb it. Subject control and object control result. ECM explains why subjects that do exist in this environment are accusative-marked. The highest argument of untensed raising predicates move to spec-TP. Some languages have V2, which can be accounted for via feature strength differences crosslinguistically. wh- questions are interesting. Usually the highest wh- item moves to spec-CP (plus I-to-C). Subjacency limits how far it can move in one step (thus defining cyclicity). Various effects (complementizer agreement, superiority, wh- islands, etc.) show this. wh- elements and quantifiers interact in interesting ways. In some languages wh- elements move before Spellout, but in some languages they move afterwards (at LF). .
Recommended publications
  • Common and Distinct Neural Substrates for Pragmatic, Semantic, and Syntactic Processing of Spoken Sentences: an Fmri Study
    Common and Distinct Neural Substrates for Pragmatic, Semantic, and Syntactic Processing of Spoken Sentences: An fMRI Study G.R.Kuperberg and P.K.McGuire Downloaded from http://mitprc.silverchair.com/jocn/article-pdf/12/2/321/1758711/089892900562138.pdf by guest on 18 May 2021 Institute of Psychiatry, London E.T.Bullmore Institute of Psychiatry, London and the University of Cambridge M.J.Brammer, S.Rabe-Hesketh, I.C.Wright, D.J.Lythgoe, S.C.R.Williams, and A.S.David Institute of Psychiatry, London Abstract & Extracting meaning from speech requires the use of difference between conditions in activation of the left-super- pragmatic, semantic, and syntactic information. A central ior-temporal gyrus for the pragmatic experiment than the question is:Does the processing of these different types of semantic/syntactic experiments; (2) a greater difference be- linguistic information have common or distinct neuroanatomi- tween conditions in activation of the right-superior and middle- cal substrates? We addressed this issue using functional temporal gyrus in the semantic experiment than in the syntactic magnetic resonance imaging (fMRI) to measure neural activity experiment; and (3) no regions activated to a greater degree in when subjects listened to spoken normal sentences contrasted the syntactic experiment than in the semantic experiment. with sentences that had either (A) pragmatical, (B) semantic These data show that, while left- and right-superior-temporal (selection restriction), or (C) syntactic (subcategorical) viola- regions may be differentially involved in processing pragmatic tions sentences. All three contrasts revealed robust activation of and lexico-semantic information within sentences, the left- the left-inferior-temporal/fusiform gyrus.
    [Show full text]
  • Modeling Subcategorization Through Co-Occurrence Outline
    Introducing LexIt Building Distributional Profiles Ongoing Work Conclusions Outline Modeling Subcategorization Through Co-occurrence 1 Introducing LexIt A Computational Lexical Resource for Italian Verbs The Project Distributional Profiles 1 2 Gabriella Lapesa ,AlessandroLenci 2 Building Distributional Profiles Pre-processing 1University of Osnabr¨uck, Institute of Cognitive Science Subcategorization Frames 2 University of Pisa, Department of Linguistics Lexical sets Selectional preferences Explorations in Syntactic Government and Subcategorisation nd University of Cambridge, 2 September 2011 3 Ongoing Work 4 Conclusions Gabriella Lapesa, Alessandro Lenci Modeling Subcategorization Through Co-occurrence 2/ 38 Introducing LexIt Introducing LexIt Building Distributional Profiles The Project Building Distributional Profiles The Project Ongoing Work Distributional Profiles Ongoing Work Distributional Profiles Conclusions Conclusions Computational approaches to argument structure LexIt: a computational lexical resource for Italian The automatic acquisition of lexical information from corpora is a longstanding research avenue in computational linguistics LexIt is a computational framework for the automatic acquisition subcategorization frames (Korhonen 2002, Schulte im Walde and exploration of corpus-based distributional profiles of Italian 2009, etc.) verbs, nouns and adjectives selectional preferences (Resnik 1993, Light & Greiff 2002, Erk et al. 2010, etc.) LexIt is publicly available through a web interface: verb classes (Merlo & Stevenson 2001, Schulte im Walde 2006, http://sesia.humnet.unipi.it/lexit/ Kipper-Schuler et al. 2008, etc.) Corpus-based information has been used to build lexical LexIt is the first large-scale resource of such type for Italian, resources aiming at characterizing the valence properties of predicates cf. VALEX for English (Korohnen et al. 2006), LexSchem for fully on distributional ground French (Messiant et al.
    [Show full text]
  • Acquiring Verb Subcategorization from Spanish Corpora
    Acquiring Verb Subcategorization from Spanish Corpora Grzegorz ChrupaÃla [email protected] Universitat de Barcelona Department of General Linguistics PhD Program “Cognitive Science and Language” Supervised by Dr. Irene Castell´onMasalles September 2003 Contents 1 Introduction 5 2 Verb Subcategorization in Linguistic Theory 7 2.1 Introduction . 7 2.2 Government-Binding and related approaches . 7 2.3 Categorial Grammar . 9 2.4 Lexical-Functional Grammar . 11 2.5 Generalized Phrase-Structure Grammar . 12 2.6 Head-Driven Phrase-Structure Grammar . 14 2.7 Discussion . 17 3 Diathesis Alternations 19 3.1 Introduction . 19 3.2 Diathesis . 19 3.2.1 Alternations . 20 3.3 Diathesis alternations in Spanish . 21 3.3.1 Change of focus . 22 3.3.2 Underspecification . 26 3.3.3 Resultative construction . 27 3.3.4 Middle construction . 27 3.3.5 Conclusions . 29 4 Verb classification 30 4.1 Introduction . 30 4.2 Semantic decomposition . 30 4.3 Levin classes . 32 4.3.1 Beth Levin’s classification . 32 4.3.2 Intersective Levin Classes . 33 4.4 Spanish: verbs of change and verbs of path . 35 4.4.1 Verbs of change . 35 4.4.2 Verbs of path . 37 4.4.3 Discussion . 39 4.5 Lexicographical databases . 40 1 4.5.1 WordNet . 40 4.5.2 VerbNet . 41 4.5.3 FrameNet . 42 5 Subcategorization Acquisition 44 5.1 Evaluation measures . 45 5.1.1 Precision, recall and the F-measure . 45 5.1.2 Types and tokens . 46 5.2 SF acquisition systems . 47 5.2.1 Raw text .
    [Show full text]
  • Features from Aspects Via the Minimalist Program to Combinatory Categorial Grammar
    Á. Gallego & D. Ott (eds.). 2015. 50 Years Later: Reflections on Chomsky’s Aspects. Cambridge, MA: MITWPL. Available for purchase at http://mitwpl.mit.edu © The Authors FEATURES FROM ASPECTS VIA THE MINIMALIST PROGRAM TO COMBINATORY CATEGORIAL GRAMMAR NEIL SMITH ANNABEL CORMACK University College London 1 Background One major contribution of Aspects (Chomsky 1965)1 was to initiate the development of a theory of syntactic features. There is no use of features, either syntactic or morphophonemic, in Chomsky’s earliest work The Morphophonemics of Modern Hebrew (1951/79); they do not appear in his monumental The Logical Structure of Linguistic Theory (1955/75); nor in Syntactic Structures (1957) (except for category labels); nor, except for phonological distinctive features, in Current Issues in Linguistic Theory (1964). In Aspects features play a crucial role. Chomsky suggests (p.214) that “We might, then, take a lexical entry to be simply a set of features, some syntactic, some phonological, some semantic”. But even here the use of features was somewhat haphazard, apparently unconstrained, and a terminological mess. In this contribution we want to trace – in broad outline – the development of a theory of morphosyntactic features and propose a fully Merge-based and parsimonious version of such a theory. Our work is intended as a contribution to Chomsky’s Minimalist Program, pursuing the same ends on the basis of some different background assumptions. These assumptions and the specific analyses derived from them have been developed using an austere version of Combinatorial Categorial Grammar, but are of wider applicability. All current theories use phonological features and ‘semantic’ features which we will largely ignore except to point out that the term ‘semantic’ is used in radically different ways to cover truth-theoretic, conceptual or encyclopaedic distinctions.
    [Show full text]
  • Glossary for Syntax in Three Dimensions (2015) © Carola Trips
    Glossary for Syntax in three Dimensions (2015) © Carola Trips A - B - C - D - E - F - G - H - I - J - K - L - M - N - O - P - Q - R - S - T - U - V - W - X - Y – Z A abstract case Abstract Case (or "Deep Case") is an abstract notion of case distinct from the morphologically marked case (cf. "case") used in Case theory (subtheories of Generative Grammar). Abstract Case is present even in languages which lack morphological case on noun phrases; it is usually assumed to be congruent with morphological case when such features are present. For example in German dative case is marked morphologically as in Ich helfe ihm ('I help him') whereas no such case marking is found in the English sentence (but Case is still there). academy In previous centuries the need for an academy dealing with linguistic matters was discussed and debated in England (on the model of the French académie française). Due to the increase in the production of grammars in the 17th and 18th centuries and the rising number of grammarians, calls for an academy to codify the English language by publishing an authoritative grammar and dictionary dwindled. acceptance Step in the process of standardization: a selected variety of a particular language must gain acceptance by a group of speakers/users who are capable of influencing other speakers. actants In valency grammar, a functional unit determined by the valency of the verb. Actants are required by the valence of the verb and are the equivalents of arguments (in generative grammar). © Carola Trips 1 active voice Term used in grammatical analysis of voice, referring to a sentence, clause or verb form where from a semantic point of view the grammatical subject is typically the actor in relation to the verb.
    [Show full text]
  • Subcategorization Semantics and the Naturalness of Verb-Frame Pairings
    University of Pennsylvania Working Papers in Linguistics Volume 4 Issue 2 Article 11 1997 Subcategorization Semantics and the Naturalness of Verb-Frame Pairings Edward Kako University of Pennsylvania Follow this and additional works at: https://repository.upenn.edu/pwpl Recommended Citation Kako, Edward (1997) "Subcategorization Semantics and the Naturalness of Verb-Frame Pairings," University of Pennsylvania Working Papers in Linguistics: Vol. 4 : Iss. 2 , Article 11. Available at: https://repository.upenn.edu/pwpl/vol4/iss2/11 This paper is posted at ScholarlyCommons. https://repository.upenn.edu/pwpl/vol4/iss2/11 For more information, please contact [email protected]. Subcategorization Semantics and the Naturalness of Verb-Frame Pairings This working paper is available in University of Pennsylvania Working Papers in Linguistics: https://repository.upenn.edu/pwpl/vol4/iss2/11 Subcategorization Semantics and the Naturalness of Verb-Frame Pairings Edward Kako 1. Introduction Do subcategorization frames have meanings independent of the verbs which appear in them? Advocates of the Projectionist position have answered “no” to this question, arguing that subcat frames serve only to satisfy the structural demands imposed upon sentences by the semantics of verbs (e.g., Chomsky, 1981; Pinker, 1989; Levin, 1993). Proponents of the Constructionist position, by contrast, have answered “yes” to this question, arguing that whether a verb and frame can peaceably cohabit in a sentence depends upon whether the two are “compatible” (e.g., Goldberg,
    [Show full text]
  • Verb Sense and Verb Subcategorization Probabilities Douglas Roland Daniel Jurafsky University of Colorado University of Colorado Department of Linguistics Dept
    To appear in Suzanne Stevenson and Paola Merlo’s volume of papers from the 1998 CUNY Sentence Processing Conference (Benjamins) Verb Sense and Verb Subcategorization Probabilities Douglas Roland Daniel Jurafsky University of Colorado University of Colorado Department of Linguistics Dept. of Linguistics, Computer Boulder, CO 80309-0295 Science, and Inst. of Cognitive Science [email protected] Boulder, CO 80309-0295 [email protected] 1 Introduction The probabilistic relation between verbs and their arguments plays an important role in psychological theories of human language processing. For example, Ford, Bresnan and Kaplan (1982) proposed that verbs like position have two lexical forms: a more preferred form that subcategorizes for three arguments (SUBJ, OBJ, PCOMP) and a less preferred form that subcategorizes for two arguments (SUBJ, OBJ). Many recent psychological experiments suggest that humans use these kinds of verb-argument preferences as an essential part of the process of sentence interpretation. (Clifton et al. 1984, Ferreira & McClure 1997, Garnsey et al. 1997, MacDonald 1994, Mitchell & Holmes 1985, Boland et al. 1990, Trueswell et al. 1993). It is not completely understood how these preferences are realized, but one possible model proposes that each lexical entry for a verb expresses a conditional probability for each potential subcategorization frame (Jurafsky 1996, Narayanan and Jurafsky 1998). Unfortunately, different methods of calculating verb subcategorization probabilities yield different results. Recent studies (Merlo 1994, Gibson et al. 1996, Roland & Jurafsky 1997) have found differences between syntactic and subcategorization frequencies computed from corpora and those computed from psychological experiments. Merlo (1994) showed that the subcategorization frequencies derived from corpus data were different from the subcategorization data derived from a variety of psychological protocols.
    [Show full text]
  • Valency Dictionary of Czech Verbs: Complex Tectogrammatical Annotation
    Valency Dictionary of Czech Verbs: Complex Tectogrammatical Annotation Marketa´ Stranˇak´ ova-Lopatk´ ova,´ Zdenekˇ Zabokrtskˇ y´ Center for Computational Linguistics Faculty of Mathematics and Physics, Charles University Malostranske´ nam.´ 25, CZ-11800 Prague, Czech Republic stranak,zabokrtsky ¡ @ckl.mff cuni.cz http://ckl.mff.cuni.cz Abstract A lexicon containing a certain kind of syntactic information about verbs is one of the crucial prerequisities for most tasks in Natural Language Processing. The goal of the project described in the paper is to create a human- and machine-readable lexicon capturing in detail valency behavior of hundreds most frequent Czech verbs. Manual annotation effort consumed at this project limits the speed of its growth on the one hand, but guarantees significantly higher data consistency than that of automatically acquired lexicons. In this paper, we outline the theoretical background on which the lexicon is based, and describe the annotation schema (lexicon data structure, annotation tools, etc.). Selected quantitative characteristics of the lexicon are presented as well. 1. Introduction The noun phrase jeho bratra [his brother] preceded by The verb is traditionally considered to be the center of no preposition can be Gen.sg or Acc.sg. The verb ptat´ the sentence, and thus the description of syntactic behav- se [to ask] allows only the former possibility. ior of verbs is a substantial task for linguists. A syntactic ¢ Syntactic analysis (considering a dependency ori- lexicon of verbs with the subcategorization information is ented formalism, syntactic analysis can be informally obviously crucial also for many tasks in Natural Language expressed as ‘determining which word depends on Processing (NLP) domain.
    [Show full text]
  • Dependency Grammar
    Formal & Computational Aspects of Dependency Grammar { Argument structure in DG { Geert-Jan M. Kruijff Computational linguistics University of the Saarland Saarbrucken¨ Germany [email protected] h i Contents First Last Prev Next J Goals Today we will have a look at two levels of argument structure, namely sub- • categorization and valency. At each level, we look at argumenthood from different viewpoints: • { Role of dependency: heads and dependents, types of relations, the meaning of relations. { Different dimensions of 'argumenthood': obligatoriness versus optionality, core versus free modification. { Instantiations of these notions in various dependency- and phrase structure- grammar frameworks. We argue that subcategorization and valency describe distinct levels of struc- • ture in a grammar. { We examplify why these levels are not isomorphic. { We develop a typological theory of abstract case to relate them. Contents First Last Prev Next J Contents Goals .................................................... 2 Subcategorization and valency .............................. 5 Subcategorization and valency are not isomorphic ....... 6 Consequences of non-isomorphy ....................... 8 Remarks on subcategorization in DG ........................ 9 Between subcategorization and surface word order ...... 12 Surface word order in DG ............................ 13 Separating dependency and linearization ............... 15 Resulting notions of subcategorization in DG ........... 17 Valency .................................................. 18
    [Show full text]
  • MITA Working Papers in Psycholinguistics, Volume 2. INSTITUTION Keio Univ., Tokyo (Japan)
    DOCUMENT RESUME ED 358 719 FL 021 283 AUTHOR Otsu, Yukio, Ed. TITLE MITA Working Papers in Psycholinguistics, Volume 2. INSTITUTION Keio Univ., Tokyo (Japan). Inst. of Cultural and Linguistic Studies. PUB DATE 89 NOTE 102p.; Volume 1 is unavailable. For selected individual papers, see ED 331 302, ED 323 759, and FL 021 286-290. Publication partly supported bya grant from the Japanese Ministry of Education and Culture and the INS Corporation. AVAILABLE FROM Institute of Cultural & Linguistics Studies, Keio University, 2-15-45 Mita, Minato-ku, Tokyo 108 Japan. PUB TYPE Collected Works General (020) EDRS PRICE MF01/PC05 Plus Postage. DESCRIPTORS College Students; *English (Second Language); English for Academic Purposes; Error Patterns; Foreign Countries; Higher Education; High School Freshmen; High Schools; High School Students; *Language Processing; Linguistic Theory; *Second Language Learning; Semanti, ; Sentence Structure; Speech Acts; Verbs; Writing (Co-position) IDENTIFIERS Clauses; Connectives (Grammar); Japan; Reflexives; Universal Grammar ABSTRACT Seven original research papers are presented. The titles and authors are as follows: "Acquisition of the Argument-Structure of Verbs" (Mika Endo); "A Note on Semantic Selection" (Yoshio Endo); "The Governing Category Parameter in Second Language" (Makiko Hirakawa); "The Use of Connectives in English Academic Papers Written by Japanese Students" (Yasuko Kanno); "A Note on English As-Clauses" (Yoshihara Kumagai); "Structure-Dependence in Second Language Acquisition" (Kazuhiro Naoi); and
    [Show full text]
  • Verb Subcategorization Frequencies: American English Corpus Data, Methodological Studies, and Cross-Corpus Comparisons
    Behavior Research Methods, Instruments, & Computers 2004, 36 (3), 432-443 Verb subcategorization frequencies: American English corpus data, methodological studies, and cross-corpus comparisons SUSANNE GAHL University of Illinois at Urbana-Champaign, Urbana, Illinois DAN JURAFSKY Stanford University, Stanford, California and DOUGLAS ROLAND University of California, San Diego, La Jolla, California Verb subcategorization frequencies (verb biases) have been widely studied in psycholinguistics and play an important role in human sentence processing. Yet available resources on subcategorization fre- quencies suffer from limited coverage, limited ecological validity, and divergent coding criteria. Prior estimates of verb transitivity, for example, vary widely with corpus size, coverage, and coding criteria. This article provides norming data for 281 verbs of interest to psycholinguistic research, sampled from a corpus of American English, along with a detailed coding manual. We examine the effect on transi- tivity bias of various coding decisions and methods of computing verb biases. The frequency of linguistic structures, such as pho- duction (Gahl & Garnsey, in press). The experimental lit- nemes or words, has long been known to affect language erature on verb biases shows that, other things being equal, processing. Increasingly, research in sentence processing sentences that conform to a verb’s bias are easier to pro- has focused on the frequencies of more complex linguis- cess than sentences that violate a verb’s bias. Indeed, this tic structures.
    [Show full text]
  • Automatic Extraction of Subcategorization from Corpora
    Automatic Extraction of Subcategorization from Corpora Ted Briscoe John Carroll Computer Laboratory Cognitive and Computing Sciences University of Cambridge University of Sussex Pembroke Street, Cambridge CB2 3QG, UK Brighton BN1 9QH, UK ejb@cl, cam. ac. uk j otto. carroll@cogs, susx. ac. uk Abstract Several substantial machine-readable subcatego- rization dictionaries exist for English, either built We describe a novel technique and imple- largely automatically from machine-readable ver- mented system for constructing a subcate- sions of conventional learners' dictionaries, or manu- gorization dictionary from textual corpora. ally by (computational) linguists (e.g. the Alvey NL Each dictionary entry encodes the relative Tools (ANLT) dictionary, Boguraev et al. (1987); frequency of occurrence of a comprehen- the COMLEX Syntax dictionary, Grishman et al. sive set of subcategorization classes for En- (1994)). Unfortunately, neither approach can yield a glish. An initial experiment, on a sample genuinely accurate or comprehensive computational of 14 verbs which exhibit multiple comple- lexicon, because both rest ultimately on the manual mentation patterns, demonstrates that the efforts of lexicographers / linguists and are, there- technique achieves accuracy comparable to fore, prone to errors of omission and commission previous approaches, which are all limited which are hard or impossible to detect automatically to a highly restricted set of subcategoriza- (e.g. Boguraev & Briscoe, 1989; see also section 3.1 tion classes. We also demonstrate that a below for an example). Furthermore, manual encod- subcategorization dictionary built with the ing is labour intensive and, therefore, it is costly to system improves the accuracy of a parser extend it to neologisms, information not currently by an appreciable amount 1.
    [Show full text]