Lexical Functional Grammar Carol Neidle, Boston University
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
The Structure of Reflexive Clauses in Michif: a Relational Grammar Approach
University of North Dakota UND Scholarly Commons Theses and Dissertations Theses, Dissertations, and Senior Projects 12-1984 The trS ucture of Reflexive Clauses in Michif: A Relational Grammar Approach Larry Lee Lovell Follow this and additional works at: https://commons.und.edu/theses Part of the Linguistics Commons Recommended Citation Lovell, Larry Lee, "The trS ucture of Reflexive Clauses in Michif: A Relational Grammar Approach" (1984). Theses and Dissertations. 903. https://commons.und.edu/theses/903 This Thesis is brought to you for free and open access by the Theses, Dissertations, and Senior Projects at UND Scholarly Commons. It has been accepted for inclusion in Theses and Dissertations by an authorized administrator of UND Scholarly Commons. For more information, please contact [email protected]. THE STRUCTURE OF REFLEXIVE CLAUSES IN MICHIF: . ' i ' A RELATIONAL GRAMMAR APPROACH ) I by Larry Lee Lovell Bachelor of icience, Middle Tennessee State University, 1982 A Thesis Submitted to the Graduate Faculty of the University of North Dakota in partial fulfillment of the requirements for the degree of Master of Arts Grand Forks, North Dakota December · 1984 This thesis submitted by Larry Lee Lovell in partial fulfillment of the requirements for the Degree of M.A. in Linguistics from the University of North Dakota is hereby approved by the Faculty Advisory Committee under whom the work has been done. 0. This thesis meets the standards for appearance and conforms to the style and format requirements of the Graduate School of the University of North Dakota, and is hereby approved. ; I ! i - ii - ' . Title THE STRUCTURE OF REFLEXIVE CLAUSES IN MICHIF: A RELATIONAL GRAMMAR APPROACH Department Linguistics Degree Master of Arts In presenting this thesis in ~artial fulfillment of the requirements for a graduate degree from the University of North Dakota, I agree that the Library of this University shall make it freely available for inspection. -
A Relational Grammar Approach to Kera Syntax Janet K
Work Papers of the Summer Institute of Linguistics, University of North Dakota Session Volume 28 Article 1 1984 A relational grammar approach to Kera syntax Janet K. Camburn SIL-UND Follow this and additional works at: https://commons.und.edu/sil-work-papers Recommended Citation Camburn, Janet K. (1984) "A relational grammar approach to Kera syntax," Work Papers of the Summer Institute of Linguistics, University of North Dakota Session: Vol. 28 , Article 1. DOI: 10.31356/silwp.vol28.01 Available at: https://commons.und.edu/sil-work-papers/vol28/iss1/1 This Thesis is brought to you for free and open access by UND Scholarly Commons. It has been accepted for inclusion in Work Papers of the Summer Institute of Linguistics, University of North Dakota Session by an authorized editor of UND Scholarly Commons. For more information, please contact [email protected]. A RELATIOIAL GRAMMAR APPROACH TO KERA SIITAX Janet K. Camburn 1 Introduction 2 Introduction to Relational Grammar 3 Kera syntax 3T1 Tone 3.2 Kera verbs 3.3 Terms and term marking 3~4 Oblique markings 4 Advancements 4.1 Passive 4~2 3 advancement 4~3 Ben-3 advancement 4.4 Summary 5 Multiple depencencies 5.1 Pro-Replacement 5.2 Equi-Erasure and Equi-Subject Union 5.3 Equi-Erasure 5~4 Equi-Subject Union 6 The proper formulation of the Stem Agreement Rule 6~1 Summary 7 Ascensions 7,1 Ascensions out of 2-host 7~2 Ascensions out of a 1-host 7.3 Ascensions and unaccusatives 8 Possessor ascension 8.1 The possessive construction in Kera 8.2 Possessor ascension to 2 8.3 Possessor ascension to 3 8.4 Summary 1 Introduction Kera is an Afroasiatic language belonging to the Chadic family of languages. -
Language Structure: Phrases “Productivity” a Property of Language • Definition – Language Is an Open System
Language Structure: Phrases “Productivity” a property of Language • Definition – Language is an open system. We can produce potentially an infinite number of different messages by combining elements differently. • Example – Words into phrases. An Example of Productivity • Human language is a communication system that bears some similarities to other animal communication systems, but is also characterized by certain unique features. (24 words) • I think that human language is a communication system that bears some similarities to other animal communication systems, but is also characterized by certain unique features, which are fascinating in and of themselves. (33 words) • I have always thought, and I have spent many years verifying, that human language is a communication system that bears some similarities to other animal communication systems, but is also characterized by certain unique features, which are fascinating in and of themselves. (42 words) • Although mainstream some people might not agree with me, I have always thought… Creating Infinite Messages • Discrete elements – Words, Phrases • Selection – Ease, Meaning, Identity • Combination – Rules of organization Models of Word reCombination 1. Word chains (Markov model) Phrase-level meaning is derived from understanding each word as it is presented in the context of immediately adjacent words. 2. Hierarchical model There are long-distant dependencies between words in a phrase, and these inform the meaning of the entire phrase. Markov Model Rule: Select and concatenate (according to meaning and what types of words should occur next to each other). bites bites bites Man over over over jumps jumps jumps house house house Markov Model • Assumption −Only adjacent words are meaningfully (and lawfully) related. -
Chapter 30 HPSG and Lexical Functional Grammar Stephen Wechsler the University of Texas Ash Asudeh University of Rochester & Carleton University
Chapter 30 HPSG and Lexical Functional Grammar Stephen Wechsler The University of Texas Ash Asudeh University of Rochester & Carleton University This chapter compares two closely related grammatical frameworks, Head-Driven Phrase Structure Grammar (HPSG) and Lexical Functional Grammar (LFG). Among the similarities: both frameworks draw a lexicalist distinction between morphology and syntax, both associate certain words with lexical argument structures, both employ semantic theories based on underspecification, and both are fully explicit and computationally implemented. The two frameworks make available many of the same representational resources. Typical differences between the analyses proffered under the two frameworks can often be traced to concomitant differ- ences of emphasis in the design orientations of their founding formulations: while HPSG’s origins emphasized the formal representation of syntactic locality condi- tions, those of LFG emphasized the formal representation of functional equivalence classes across grammatical structures. Our comparison of the two theories includes a point by point syntactic comparison, after which we turn to an exposition ofGlue Semantics, a theory of semantic composition closely associated with LFG. 1 Introduction Head-Driven Phrase Structure Grammar is similar in many respects to its sister framework, Lexical Functional Grammar or LFG (Bresnan et al. 2016; Dalrymple et al. 2019). Both HPSG and LFG are lexicalist frameworks in the sense that they distinguish between the morphological system that creates words and the syn- tax proper that combines those fully inflected words into phrases and sentences. Stephen Wechsler & Ash Asudeh. 2021. HPSG and Lexical Functional Gram- mar. In Stefan Müller, Anne Abeillé, Robert D. Borsley & Jean- Pierre Koenig (eds.), Head-Driven Phrase Structure Grammar: The handbook. -
Lexical-Functional Grammar and Order-Free Semantic Composition
COLING 82, J. Horeck~(eel) North-HoOandPubllshi~ Company O A~deml~ 1982 Lexical-Functional Grammar and Order-Free Semantic Composition Per-Kristian Halvorsen Norwegian Research Council's Computing Center for the Humanities and Center for Cognitive Science, MIT This paper summarizes the extension of the theory of lexical-functional grammar to include a formal, model-theoretic, semantics. The algorithmic specification of the semantic interpretation procedures is order-free which distinguishes the system from other theories providing model-theoretic interpretation for natural language. Attention is focused on the computational advantages of a semantic interpretation system that takes as its input functional structures as opposed to syntactic surface-structures. A pressing problem for computational linguistics is the development of linguistic theories which are supported by strong independent linguistic argumentation, and which can, simultaneously, serve as a basis for efficient implementations in language processing systems. Linguistic theories with these properties make it possible for computational implementations to build directly on the work of linguists both in the area of grammar-writing, and in the area of theory development (cf. universal conditions on anaphoric binding, filler-gap dependencies etc.). Lexical-functional grammar (LFG) is a linguistic theory which has been developed with equal attention being paid to theoretical linguistic and computational processing considerations (Kaplan & Bresnan 1981). The linguistic theory has ample and broad motivation (vide the papers in Bresnan 1982), and it is transparently implementable as a syntactic parsing system (Kaplan & Halvorsen forthcoming). LFG takes grammatical relations to be of primary importance (as opposed to the transformational theory where grammatical functions play a subsidiary role). -
Can Participate in Syntactic Operations; the Law Means That Any Oblique Which Appears in a Sentence Must Bear That Oblique Relation in the Ini Tial Level
0SU WPL # 26 (1982) 93-101. A Note on the Oblique Law* Brian D. Joseph Perlmutter and Postal (1978a: 15) posit the following "law" within the framework of Relational Grammar : (1) We say that Bis a Ci arc if Bis an arc one of whose coordinates is Ci. Then: if A is an oblique arc , A is a C1 arc. This law is known as the Oblique Law because it constrains the extent to which oblique nominals, i.e. those bearing the nonterm core grammatical rel ations1 such as benefactive, temporal, circumstantial, locative, etc. can participate in syntactic operations; the law means that any oblique which appears in a sentence must bear that oblique relation in the ini tial level. The original intent of this law as indicated by Perlmutter and Postal ' s discussion (p. 15) of it, was to rule out advancements or demotions2 to any oblique grammatical relation. However, as it is currently formulated , it is more general than that and rules out ascensions (i. e . raisings) to an oblique relation as well. The purpose of this note is to show that at best, only the more restricted interpretation of the Oblique Law is valid, because there is a construction in Modern Greek which provides a counter- example to the broader interpretation, as well as other facts which might bear on even the more narrow version of the law. The Greek construction in question is the one called Raising-to-Oblique in Joseph (1979), and involves the circumstantial preposition me ' with' .3 Me can govern simple nominal complements, as in (2), or clausalcomplements, as in (3) : (2) me t6so 66rivo, aen borusa na aulepso with so-much noise/ACC not could/SG VBL PART work/1 SG ' With so much noise, I couldn't work.' (3) me to na stekete eki etsi with NOMINALIZER VBL PART stand/3 S~ there thus o Yanis, tlen borusa na dulepso John/NOM ' With John standing there like that, I couldn' t work.' The evidence against the Oblique Law comes from a construction which is a variant of (3), in which the subject of the clausal object of me is raised to become itself the object of me. -
Psychoacoustics, Speech Perception, Language Structure and Neurolinguistics Hearing Acuity Absolute Auditory Threshold Constant
David House: Psychoacoustics, speech perception, 2018.01.25 language structure and neurolinguistics Hearing acuity Psychoacoustics, speech perception, language structure • Sensitive for sounds from 20 to 20 000 Hz and neurolinguistics • Greatest sensitivity between 1000-6000 Hz • Non-linear perception of frequency intervals David House – E.g. octaves • 100Hz - 200Hz - 400Hz - 800Hz - 1600Hz – 100Hz - 800Hz perceived as a large difference – 3100Hz - 3800 Hz perceived as a small difference Absolute auditory threshold Demo: SPL (Sound pressure level) dB • Decreasing noise levels – 6 dB steps, 10 steps, 2* – 3 dB steps, 15 steps, 2* – 1 dB steps, 20 steps, 2* Constant loudness levels in phons Demo: SPL and loudness (phons) • 50-100-200-400-800-1600-3200-6400 Hz – 1: constant SPL 40 dB, 2* – 2: constant 40 phons, 2* 1 David House: Psychoacoustics, speech perception, 2018.01.25 language structure and neurolinguistics Critical bands • Bandwidth increases with frequency – 200 Hz (critical bandwidth 50 Hz) – 800 Hz (critical bandwidth 80 Hz) – 3200 Hz (critical bandwidth 200 Hz) Critical bands demo Effects of masking • Fm=200 Hz (critical bandwidth 50 Hz) – B= 300,204,141,99,70,49,35,25,17,12 Hz • Fm=800 Hz (critical bandwidth 80 Hz) – B=816,566,396,279,197,139,98,69,49,35 Hz • Fm=3200 Hz (critical bandwidth 200 Hz) – B=2263,1585,1115,786,555,392,277,196,139,98 Hz Effects of masking Holistic vs. analytic listening • Low frequencies more effectively mask • Demo 1: audible harmonics (1-5) high frequencies • Demo 2: melody with harmonics • Demo: how -
Phrase Structure Rules, Tree Rewriting, and Other Sources of Recursion Structure Within the NP
Introduction to Transformational Grammar, LINGUIST 601 September 14, 2006 Phrase Structure Rules, Tree Rewriting, and other sources of Recursion Structure within the NP 1 Trees (1) a tree for ‘the brown fox sings’ A ¨¨HH ¨¨ HH B C ¨H ¨¨ HH sings D E the ¨¨HH F G brown fox • Linguistic trees have nodes. The nodes in (1) are A, B, C, D, E, F, and G. • There are two kinds of nodes: internal nodes and terminal nodes. The internal nodes in (1) are A, B, and E. The terminal nodes are C, D, F, and G. Terminal nodes are so called because they are not expanded into anything further. The tree ends there. Terminal nodes are also called leaf nodes. The leaves of (1) are really the words that constitute the sentence ‘the brown fox sings’ i.e. ‘the’, ‘brown’, ‘fox’, and ‘sings’. (2) a. A set of nodes form a constituent iff they are exhaustively dominated by a common node. b. X is a constituent of Y iff X is dominated by Y. c. X is an immediate constituent of Y iff X is immediately dominated by Y. Notions such as subject, object, prepositional object etc. can be defined structurally. So a subject is the NP immediately dominated by S and an object is an NP immediately dominated by VP etc. (3) a. If a node X immediately dominates a node Y, then X is the mother of Y, and Y is the daughter of X. b. A set of nodes are sisters if they are all immediately dominated by the same (mother) node. -
F I Rrrrrrrrr N Rr
fI 1 1 3 rrrrrrrrr1822 00070 n rr8354 UNIVERSITY OF CALIFORNIA San Diego The Relational Structure of Turkish Syntax A dissertation submitted in partial satisfaction of the requirements for the degree Doctor of Philosophy in Linguistics by Inci Zilhra Ozkaragoz Committee in charge: Professor Sandra Chung, Chairperson Professor Edward Klima Professor Margaret Langdon Professor David Jordan Professor Michael Meeker 1986 Copyright 1986 by Inci• Zuhra•• Ozkaragoz•• - The dissertation of Inci. Zuhra.. .Ozkaragoz .. is approved, and it is acceptable in quality and form for publication on microfilm: I /.- &l a--,_ .. /<Zx-v-·/ ·· Chairperson University of California, San Diego 1986 iii Table of Contents Page List of Abbreviations . vii Acknowledgements ................................ viii Vita, Publications and Fields of Study .......... x Abstract . xi Introduction • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • • 1 1 Relational Grammar ........................... 6 1 • 1 Introduction .............................. 6 1 . 2 Clause Structure .......................... 7 1 • 3 Typology of Strata and Some Defined Concepts ......................... 1 1 1 • 4 Well-formedness Conditions ................ 1 3 2 Aspects of Turkish Grammar ................... 1 4 2. 1 Introduction .............................. 1 4 2.2 Word Order ................................ 1 4 2.3 Nominal Morphology ........................ 16 2. 3. 1 Case Marking and Articles ................. 1 6 2.3.2 Pronouns .................................. 29 2.4 Verbal -
Ccgbank: a Corpus of CCG Derivations and Dependency Structures Extracted from the Penn Treebank
CCGbank: A Corpus of CCG Derivations and Dependency Structures Extracted from the Penn Treebank ∗ Julia Hockenmaier University of Pennsylvania ∗∗ Mark Steedman University of Edinburgh This article presents an algorithm for translating the Penn Treebank into a corpus of Combina- tory Categorial Grammar (CCG) derivations augmented with local and long-range word–word dependencies. The resulting corpus, CCGbank, includes 99.4% of the sentences in the Penn Treebank. It is available from the Linguistic Data Consortium, and has been used to train wide- coverage statistical parsers that obtain state-of-the-art rates of dependency recovery. In order to obtain linguistically adequate CCG analyses, and to eliminate noise and incon- sistencies in the original annotation, an extensive analysis of the constructions and annotations in the Penn Treebank was called for, and a substantial number of changes to the Treebank were necessary. We discuss the implications of our findings for the extraction of other linguistically expressive grammars from the Treebank, and for the design of future treebanks. 1. Introduction In order to understand a newspaper article, or any other piece of text, it is necessary to construct a representation of its meaning that is amenable to some form of inference. This requires a syntactic representation which is transparent to the underlying seman- tics, making the local and long-range dependencies between heads, arguments, and modifiers explicit. It also requires a grammar that has sufficient coverage to deal with the vocabulary and the full range of constructions that arise in free text, together with a parsing model that can identify the correct analysis among the many alternatives that such a wide-coverage grammar will generate even for the simplest sentences. -
The Minimalist Program 1 1 the Minimalist Program
The Minimalist Program 1 1 The Minimalist Program Introduction It is my opinion that the implications of the Minimalist Program (MP) are more radical than generally supposed. I do not believe that the main thrust of MP is technical; whether to move features or categories for example. MP suggests that UG has a very different look from the standard picture offered by GB-based theories. This book tries to make good on this claim by outlining an approach to grammar based on one version of MP. I stress at the outset the qualifier “version.” Minimalism is not a theory but a program animated by certain kinds of methodological and substantive regulative ideals. These ideals are reflected in more concrete principles which are in turn used in minimalist models to analyze specific empirical phenomena. What follows is but one way of articulating the MP credo. I hope to convince you that this version spawns grammatical accounts that have a theoretically interesting structure and a fair degree of empirical support. The task, however, is doubly difficult. First, it is unclear what the content of these precepts is. Second, there is a non-negligible distance between the content of such precepts and its formal realization in specific grammatical principles and analyses. The immediate task is to approach the first hurdle and report what I take the precepts and principles of MP to be.1 1 Principles-Parameters and Minimalism MP is many things to many researchers. To my mind it grows out of the per- ceived success of the principles and parameters (P&P) approach to grammatical competence. -
LSA 112 Notes on Basic Phrase Structure
LSA 112 Notes on basic phrase structure Ida Toivonen 1 Syntactic phrases A syntactic phrase is a syntactic unit that consists of one or more words. We’ll assume here that the major word classes project phrases. By ‘major’ word classes, we mean the lexical (N, V, A, Adv, P) as opposed to functional word classes (Det, Conj, Comp, Aux). Each syntactic phrase has a head. • The head of a noun phrase (NP) is a noun. • The head of a verb phrase (VP) is a verb. • The head of an adjective phrase (AP) is an adjective. • The head of an adverb phrase (AdvP) is an adverb. • The head of a prepositional phrase (PP) is a preposition. 2 Constituents The smallest syntactic unit of a sentence is a word. Larger units are phrases. A sentence (S) is of course also a unit. These units are constituents of a sentence. Each word is a constituent, each phrase is a constituent, and the sentence itself (S) is a constituent. (1) Mary saw the cat. (2) [[[Mary]] [[saw][[the][cat]]]] (3) [S[NP [N Mary]] [VP [V saw][NP [Det the][N cat]]]] Sometimes one might want to show a simplified representation. For example, one may decide that it is unnecessary (in a given situation) to bracket the individual words. (Note that Mary is the only word in the NP.) (4) [S[NP Mary] [VP saw [NP the cat]]] An alternative notation to bracketing notation is the syntactic tree (phrase marker, phrase struc- ture tree). 1 (5) S NP VP N V NP Mary saw det N the cat Phrase markers represent the hierarchical structure and linear order of sentences.