What Is Linguistics? Part II
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Lisa Pearl LING200, Summer Session I, 2004 Syntax – Sentence Structure
Lisa Pearl LING200, Summer Session I, 2004 Syntax – Sentence Structure I. Syntax A. Gives us the ability to say something new with old words. Though you have seen all the words that I’m using in this sentence before, it’s likely you haven’t seen them put together in exactly this way before. But you’re perfectly capable of understanding the sentence, nonetheless. A. Structure: how to put words together. A. Grammatical structure: a way to put words together which is allowed in the language. a. I often paint my nails green or blue. a. *Often green my nails I blue paint or. II. Syntactic Categories A. How to tell which category a word belongs to: look at the type of meaning the word has, the type of affixes which can adjoin to the word, and the environments you find the word in. A. Meaning: tell lexical category from nonlexical category. a. lexical – meaning is fairly easy to paraphrase i. vampire ≈ “night time creature that drinks blood” i. dark ≈ “without much light” i. drink ≈ “ingest a liquid” a. nonlexical – meaning is much harder to pin down i. and ≈ ?? i. the ≈ ?? i. must ≈ ?? a. lexical categories: Noun, Verb, Adjective, Adverb, Preposition i. N: blood, truth, life, beauty (≈ entities, objects) i. V: bleed, tell, live, remain (≈ actions, sensations, states) i. A: gory, truthful, riveting, beautiful (≈ property/attribute of N) 1. gory scene 1. riveting tale i. Adv: sharply, truthfully, loudly, beautifully (≈ property/attribute of V) 1. speak sharply 1. dance beautifully i. P: to, from, inside, around, under (≈ indicates path or position w.r.t N) 1. -
Logophoricity in Finnish
Open Linguistics 2018; 4: 630–656 Research Article Elsi Kaiser* Effects of perspective-taking on pronominal reference to humans and animals: Logophoricity in Finnish https://doi.org/10.1515/opli-2018-0031 Received December 19, 2017; accepted August 28, 2018 Abstract: This paper investigates the logophoric pronoun system of Finnish, with a focus on reference to animals, to further our understanding of the linguistic representation of non-human animals, how perspective-taking is signaled linguistically, and how this relates to features such as [+/-HUMAN]. In contexts where animals are grammatically [-HUMAN] but conceptualized as the perspectival center (whose thoughts, speech or mental state is being reported), can they be referred to with logophoric pronouns? Colloquial Finnish is claimed to have a logophoric pronoun which has the same form as the human-referring pronoun of standard Finnish, hän (she/he). This allows us to test whether a pronoun that may at first blush seem featurally specified to seek [+HUMAN] referents can be used for [-HUMAN] referents when they are logophoric. I used corpus data to compare the claim that hän is logophoric in both standard and colloquial Finnish vs. the claim that the two registers have different logophoric systems. I argue for a unified system where hän is logophoric in both registers, and moreover can be used for logophoric [-HUMAN] referents in both colloquial and standard Finnish. Thus, on its logophoric use, hän does not require its referent to be [+HUMAN]. Keywords: Finnish, logophoric pronouns, logophoricity, anti-logophoricity, animacy, non-human animals, perspective-taking, corpus 1 Introduction A key aspect of being human is our ability to think and reason about our own mental states as well as those of others, and to recognize that others’ perspectives, knowledge or mental states are distinct from our own, an ability known as Theory of Mind (term due to Premack & Woodruff 1978). -
II Levels of Language
II Levels of language 1 Phonetics and phonology 1.1 Characterising articulations 1.1.1 Consonants 1.1.2 Vowels 1.2 Phonotactics 1.3 Syllable structure 1.4 Prosody 1.5 Writing and sound 2 Morphology 2.1 Word, morpheme and allomorph 2.1.1 Various types of morphemes 2.2 Word classes 2.3 Inflectional morphology 2.3.1 Other types of inflection 2.3.2 Status of inflectional morphology 2.4 Derivational morphology 2.4.1 Types of word formation 2.4.2 Further issues in word formation 2.4.3 The mixed lexicon 2.4.4 Phonological processes in word formation 3 Lexicology 3.1 Awareness of the lexicon 3.2 Terms and distinctions 3.3 Word fields 3.4 Lexicological processes in English 3.5 Questions of style 4 Syntax 4.1 The nature of linguistic theory 4.2 Why analyse sentence structure? 4.2.1 Acquisition of syntax 4.2.2 Sentence production 4.3 The structure of clauses and sentences 4.3.1 Form and function 4.3.2 Arguments and complements 4.3.3 Thematic roles in sentences 4.3.4 Traces 4.3.5 Empty categories 4.3.6 Similarities in patterning Raymond Hickey Levels of language Page 2 of 115 4.4 Sentence analysis 4.4.1 Phrase structure grammar 4.4.2 The concept of ‘generation’ 4.4.3 Surface ambiguity 4.4.4 Impossible sentences 4.5 The study of syntax 4.5.1 The early model of generative grammar 4.5.2 The standard theory 4.5.3 EST and REST 4.5.4 X-bar theory 4.5.5 Government and binding theory 4.5.6 Universal grammar 4.5.7 Modular organisation of language 4.5.8 The minimalist program 5 Semantics 5.1 The meaning of ‘meaning’ 5.1.1 Presupposition and entailment 5.2 -
Making Logical Form Type-Logical Glue Semantics for Minimalist Syntax
Making Logical Form type-logical Glue Semantics for Minimalist syntax Matthew Gotham University of Oslo UiO Forum for Theoretical Linguistics 12 October 2016 Slides available at <http://folk.uio.no/matthegg/research#talks> Matthew Gotham (Oslo) Making LF type-logical FTL, 12.10.2016 1 / 63 What this talk is about Slides available at <http://folk.uio.no/matthegg/research#talks> An implementation of Glue Semantics —an approach that treats the syntax-semantics interface as deduction in a type logic— for Minimalist syntax, i.e. syntactic theories in the ST!EST!REST!GB!...‘Chomskyan’ tradition. Q How Minimalist, as opposed to (say) GB-ish? A Not particularly, but the factoring together of subcategorization and structure building (in the mechanism of feature-checking) is, if not crucial to this analysis, then certainly useful. and a comparison of this approach with more mainstream approaches to the syntax-semantics interface. Matthew Gotham (Oslo) Making LF type-logical FTL, 12.10.2016 2 / 63 Outline 1 The mainstream approach 2 A fast introduction to Glue Semantics 3 Implementation in Minimalism The form of syntactic theory assumed The connection to Glue 4 Comparison with the mainstream approach Interpreting (overt) movement Problems with the mainstream approach Glue analysis Nested DPs Scope islands Matthew Gotham (Oslo) Making LF type-logical FTL, 12.10.2016 3 / 63 The mainstream approach How semantics tends to be done for broadly GB/P&P/Minimalist syntax Aer Heim & Kratzer (1998) Syntax produces structures that are interpreted recursively -
A More Perfect Tree-Building Machine 1. Refining Our Phrase-Structure
Intro to Linguistics Syntax 2: A more perfect Tree-Building Machine 1. Refining our phrase-structure rules Rules we have so far (similar to the worksheet, compare with previous handout): a. S → NP VP b. NP → (D) AdjP* N PP* c. VP → (AdvP) V (NP) PP* d. PP → P NP e. AP --> (Deg) A f. AdvP → (Deg) Adv More about Noun Phrases Last week I drew trees for NPs that were different from those in the book for the destruction of the city • What is the difference between the two theories of NP? Which one corresponds to the rule (b)? • What does the left theory claim that the right theory does not claim? • Which is closer to reality? How did can we test it? On a related note: how many meanings does one have? 1) John saw a large pink plastic balloon, and I saw one, too. Let us improve our NP-rule: b. CP = Complementizer Phrase Sometimes we get a sentence inside another sentence: Tenors from Odessa know that sopranos from Boston sing songs of glory. We need to have a new rule to build a sentence that can serve as a complement of a verb, where C complementizer is “that” , “whether”, “if” or it could be silent (unpronounced). g. CP → C S We should also improve our rule (c) to be like the one below. c. VP → (AdvP) V (NP) PP* (CP) (says CP can co-occur with NP or PP) Practice: Can you draw all the trees in (2) using the amended rules? 2) a. Residents of Boston hear sopranos of renown at dusk. -
Proceedings of the LFG 00 Conference University of California, Berkeley Editors: Miriam Butt and Tracy Holloway King 2000 CSLI P
Proceedings of the LFG 00 Conference University of California, Berkeley Editors: Miriam Butt and Tracy Holloway King 2000 CSLI Publications ISSN 1098-6782 Editors' Note This year's conference was held as part of a larger Berkeley Formal Grammar Conference, which included a day of workshops and the annual HPSG conference. The program committee for LFG'00 were Rachel Nordlinger and Chris Manning. We would like to thank them for putting together the program that gave rise to this collection of papers. We would also like to thank all those who contributed to this conference, especially the local organizing committee, namely Andreas Kathol, and our reviewers, without whom the conference would not have been possible. Miriam Butt and Tracy Holloway King Functional Identity and Resource-Sensitivity in Control Ash Asudeh Stanford University and Xerox PARC Proceedings of the LFG00 Conference University of California, Berkeley Miriam Butt and Tracy Holloway King (Editors) 2000 CSLI Publications http://csli-publications.stanford.edu/ 2 1 Introduction1 Glue semantics provides a semantics for Lexical Functional Grammar (LFG) that is expressed using linear logic (Girard, 1987; Dalrymple, 1999) and provides an interpretation for the f(unctional)-structure level of syntactic representation, connecting it to the level of s(emantic)-structure in LFG’s parallel projection archi- tecture (Kaplan, 1987, 1995). Due to its use of linear logic for meaning assembly, Glue is resource-sensitive: semantic resources contributed by lexical entries and resulting f-structures must each be used in a successful proof exactly once. In this paper, I will examine the tension between a resource-sensitive semantics which interprets f-structures and structure-sharing in f-structures as expressed by functional control resulting from lexical functional identity equations. -
Toward a Shared Syntax for Shifted Indexicals and Logophoric Pronouns
Toward a Shared Syntax for Shifted Indexicals and Logophoric Pronouns Mark Baker Rutgers University April 2018 Abstract: I argue that indexical shift is more like logophoricity and complementizer agreement than most previous semantic accounts would have it. In particular, there is evidence of a syntactic requirement at work, such that the antecedent of a shifted “I” must be a superordinate subject, just as the antecedent of a logophoric pronoun or the goal of complementizer agreement must be. I take this to be evidence that the antecedent enters into a syntactic control relationship with a null operator in all three constructions. Comparative data comes from Magahi and Sakha (for indexical shift), Yoruba (for logophoric pronouns), and Lubukusu (for complementizer agreement). 1. Introduction Having had an office next to Lisa Travis’s for 12 formative years, I learned many things from her that still influence my thinking. One is her example of taking semantic notions, such as aspect and event roles, and finding ways to implement them in syntactic structure, so as to advance the study of less familiar languages and topics.1 In that spirit, I offer here some thoughts about how logophoricity and indexical shift, topics often discussed from a more or less semantic point of view, might have syntactic underpinnings—and indeed, the same syntactic underpinnings. On an impressionistic level, it would not seem too surprising for logophoricity and indexical shift to have a common syntactic infrastructure. Canonical logophoricity as it is found in various West African languages involves using a special pronoun inside the finite CP complement of a verb to refer to the subject of that verb. -
RELATIONAL NOUNS, PRONOUNS, and Resumptionw Relational Nouns, Such As Neighbour, Mother, and Rumour, Present a Challenge to Synt
Linguistics and Philosophy (2005) 28:375–446 Ó Springer 2005 DOI 10.1007/s10988-005-2656-7 ASH ASUDEH RELATIONAL NOUNS, PRONOUNS, AND RESUMPTIONw ABSTRACT. This paper presents a variable-free analysis of relational nouns in Glue Semantics, within a Lexical Functional Grammar (LFG) architecture. Rela- tional nouns and resumptive pronouns are bound using the usual binding mecha- nisms of LFG. Special attention is paid to the bound readings of relational nouns, how these interact with genitives and obliques, and their behaviour with respect to scope, crossover and reconstruction. I consider a puzzle that arises regarding rela- tional nouns and resumptive pronouns, given that relational nouns can have bound readings and resumptive pronouns are just a specific instance of bound pronouns. The puzzle is: why is it impossible for bound implicit arguments of relational nouns to be resumptive? The puzzle is highlighted by a well-known variety of variable-free semantics, where pronouns and relational noun phrases are identical both in category and (base) type. I show that the puzzle also arises for an established variable-based theory. I present an analysis of resumptive pronouns that crucially treats resumptives in terms of the resource logic linear logic that underlies Glue Semantics: a resumptive pronoun is a perfectly ordinary pronoun that constitutes a surplus resource; this surplus resource requires the presence of a resumptive-licensing resource consumer, a manager resource. Manager resources properly distinguish between resumptive pronouns and bound relational nouns based on differences between them at the level of semantic structure. The resumptive puzzle is thus solved. The paper closes by considering the solution in light of the hypothesis of direct compositionality. -
A Noun Phrase Parser of English Atro Voutilainen H Elsinki
A Noun Phrase Parser of English Atro Voutilainen H elsinki A bstract An accurate rule-based noun phrase parser of English is described. Special attention is given to the linguistic description. A report on a performance test concludes the paper. 1. Introduction 1.1 Motivation. A noun phrase parser is useful for several purposes, e.g. for index term generation in an information retrieval application; for the extraction of collocational knowledge from large corpora for the development of computational tools for language analysis; for providing a shallow but accurately analysed input for a more ambitious parsing system; for the discovery of translation units, and so on. Actually, the present noun phrase parser is already used in a noun phrase extractor called NPtool (Voutilainen 1993). 1.2. Constraint Grammar. The present system is based on the Constraint Grammar framework originally proposed by Karlsson (1990). A few characteristics of this framework are in order. • The linguistic representation is based on surface-oriented morphosyntactic tags that can encode dependency-oriented functional relations between words. • Parsing is reductionistic. All conventional analyses are provided as alternatives to each word by a context-free lookup mechanism, typically a morphological analyser. The parser itself seeks to discard all and only the contextually illegitimate alternative readings. What 'survives' is the parse.• • The system is modular and sequential. For instance, a grammar for the resolution of morphological (or part-of-speech) ambiguities is applied, before a syntactic module is used to introduce and then resolve syntactic ambiguities. 301 Proceedings of NODALIDA 1993, pages 301-310 • The parsing description is based on linguistic generalisations rather than probabilities. -
Personal Pronouns, Pronoun-Antecedent Agreement, and Vague Or Unclear Pronoun References
Personal Pronouns, Pronoun-Antecedent Agreement, and Vague or Unclear Pronoun References PERSONAL PRONOUNS Personal pronouns are pronouns that are used to refer to specific individuals or things. Personal pronouns can be singular or plural, and can refer to someone in the first, second, or third person. First person is used when the speaker or narrator is identifying himself or herself. Second person is used when the speaker or narrator is directly addressing another person who is present. Third person is used when the speaker or narrator is referring to a person who is not present or to anything other than a person, e.g., a boat, a university, a theory. First-, second-, and third-person personal pronouns can all be singular or plural. Also, all of them can be nominative (the subject of a verb), objective (the object of a verb or preposition), or possessive. Personal pronouns tend to change form as they change number and function. Singular Plural 1st person I, me, my, mine We, us, our, ours 2nd person you, you, your, yours you, you, your, yours she, her, her, hers 3rd person he, him, his, his they, them, their, theirs it, it, its Most academic writing uses third-person personal pronouns exclusively and avoids first- and second-person personal pronouns. MORE . PRONOUN-ANTECEDENT AGREEMENT A personal pronoun takes the place of a noun. An antecedent is the word, phrase, or clause to which a pronoun refers. In all of the following examples, the antecedent is in bold and the pronoun is italicized: The teacher forgot her book. -
Chapter 22 Semantics Jean-Pierre Koenig University at Buffalo Frank Richter Goethe Universität Frankfurt
Chapter 22 Semantics Jean-Pierre Koenig University at Buffalo Frank Richter Goethe Universität Frankfurt This chapter discusses the integration of theories of semantic representations into HPSG. It focuses on those aspects that are specific to HPSG and, in particular, re- cent approaches that make use of underspecified semantic representations, as they are quite unique to HPSG. 1 Introduction A semantic level of description is more integrated into the architecture of HPSG than in many frameworks (although, in the last couple of decades, the integra- tion of syntax and semantics has become tighter overall; see Heim & Kratzer 1998 for Mainstream Generative Grammar1, for example). Every node in a syntactic tree includes all appropriate levels of structure, phonology, syntax, semantics, and pragmatics so that local interaction between all these levels is in principle possible within the HPSG architecture. The architecture of HPSG thus follows the spirit of the rule-to-rule approach advocated in Bach (1976) and more specifi- cally Klein & Sag (1985) to have every syntactic operation matched by a semantic operation (the latter, of course, follows the Categorial Grammar lead, broadly speaking; Ajdukiewicz 1935; Pollard 1984; Steedman 2000). But, as we shall see, only the spirit of the rule-to-rule approach is adhered to, as there can be more 1We follow Culicover & Jackendoff (2005: 3) in using the term Mainstream Generative Grammar (MGG) to refer to work in Government & Binding or Minimalism. Jean-Pierre Koenig & Frank Richter. 2021. Semantics. In Stefan Müller, Anne Abeillé, Robert D. Borsley & Jean- Pierre Koenig (eds.), Head-Driven Phrase Structure Grammar: The handbook. Prepublished version. -
Language Structure: Phrases “Productivity” a Property of Language • Definition – Language Is an Open System
Language Structure: Phrases “Productivity” a property of Language • Definition – Language is an open system. We can produce potentially an infinite number of different messages by combining elements differently. • Example – Words into phrases. An Example of Productivity • Human language is a communication system that bears some similarities to other animal communication systems, but is also characterized by certain unique features. (24 words) • I think that human language is a communication system that bears some similarities to other animal communication systems, but is also characterized by certain unique features, which are fascinating in and of themselves. (33 words) • I have always thought, and I have spent many years verifying, that human language is a communication system that bears some similarities to other animal communication systems, but is also characterized by certain unique features, which are fascinating in and of themselves. (42 words) • Although mainstream some people might not agree with me, I have always thought… Creating Infinite Messages • Discrete elements – Words, Phrases • Selection – Ease, Meaning, Identity • Combination – Rules of organization Models of Word reCombination 1. Word chains (Markov model) Phrase-level meaning is derived from understanding each word as it is presented in the context of immediately adjacent words. 2. Hierarchical model There are long-distant dependencies between words in a phrase, and these inform the meaning of the entire phrase. Markov Model Rule: Select and concatenate (according to meaning and what types of words should occur next to each other). bites bites bites Man over over over jumps jumps jumps house house house Markov Model • Assumption −Only adjacent words are meaningfully (and lawfully) related.