LEXICAL-FUNCTIONAL JOAN BRESNAN, ASH ASUDEH, IDA TOIVONEN AND STEPHEN WECHSLER SECOND EDITION

Lexical-Functional Syntax Blackwell Textbooks in Linguistics

The books included in this series provide comprehensive accounts of some of the most central and most rapidly developing areas of research in linguistics. Intended primarily for introductory and post-introductory students, they include exercises, discussion points and suggestions for further reading.

1. Liliane Haegeman, Introduction to Government and Binding Theory (Sec- ond Edition) 2. Andrew Spencer, Morphological Theory 3. Helen Goodluck, Language Acquisition 4. Ronald Wardhaugh and Janet M. Fuller, An Introduction to Sociolinguistics (Seventh Edition) 5. Martin Atkinson, Children’s Syntax 6. Diane Blakemore, Understanding Utterances 7. Michael Kenstowicz, Phonology in Generative Grammar 8. Deborah Schiffrin, Approaches to Discourse 9. John Clark, Colin Yallop, and Janet Fletcher, An Introduction to Phonetics and Phonology (Third Edition) 10. Natsuko Tsujimura, An Introduction to Japanese Linguistics (Third Edition) 11. Robert D. Borsley, Modern Phrase Structure Grammar 12. Nigel Fabb, Linguistics and Literature 13. Irene Heim and Angelika Kratzer, Semantics in Generative Grammar 14. Liliane Haegeman and Jacqueline Gueron,´ English Grammar: A Generative Perspective 15. Stephen Crain and Diane Lillo-Martin, An Introduction to Linguistic Theory and Language Acquisition 16. Joan Bresnan, Ash Asudeh, Ida Toivonen and Stephen Wechsler, Lexical- Functional Syntax (Second Edition) 17. Barbara A. Fennell, A History of English: A Sociolinguistic Approach 18. Henry Rogers, Writing Systems: A Linguistic Approach 19. Benjamin W. Fortson IV, Indo-European Language and Culture: An Intro- duction (Second Edition) 20. Liliane Haegeman, Thinking Syntactically: A Guide to Argumentation and Analysis 21. Mark Hale, Historical Linguistics: Theory and Method 22. Henning Reetz and Allard Jongman, Phonetics: Transcription, Production, Acoustics and Perception 23. Bruce Hayes, Introductory Phonology 24. Betty J. Birner, Introduction to Pragmatics 25. Ronald Wardhaugh and Janet M. Fuller, An Introduction to Sociolinguistics Lexical-Functional Syntax

Second Edition

Joan Bresnan Stanford University

Ash Asudeh University of Oxford & Carleton University

Ida Toivonen Carleton University

Stephen Wechsler The University of Texas at Austin This second edition first published 2016

© 2016 Joan Bresnan, Ash Asudeh, Ida Toivonen, and Stephen Wechsler

Edition History: Blackwell Publishers Ltd (1e, 2001)

Registered Office John Wiley & Sons, Ltd, The Atrium, Southern Gate, Chichester, West Sussex, PO19 8SQ, UK

Editorial Offices 350 Main Street, Malden, MA 02148-5020, USA 9600 Garsington Road, Oxford, OX4 2DQ, UK The Atrium, Southern Gate, Chichester, West Sussex, PO19 8SQ, UK

For details of our global editorial offices, for customer services, and for information about how to applyfor permission to reuse the copyright material in this book please see our website at www.wiley.com/wiley-blackwell.

The right of Joan Bresnan, Ash Asudeh, Ida Toivonen, and Stephen Wechsler to be identified as the authors of this work has been asserted in accordance with the UK Copyright, Designs and Patents Act 1988.

All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording or otherwise, except as permitted by the UK Copyright, Designs and Patents Act 1988, without the prior permission of the publisher.

Wiley also publishes its books in a variety of electronic formats. Some content that appears in print may not be available in electronic books.

Designations used by companies to distinguish their products are often claimed as trademarks. All brand names and product names used in this book are trade names, service marks, trademarks or registered trademarks of their respective owners. The publisher is not associated with any product or vendor mentioned in this book.

Limit of Liability/Disclaimer of Warranty: While the publisher and authors have used their best efforts in preparing this book, they make no representations or warranties with respect to the accuracy or completeness of the contents of this book and specifically disclaim any implied warranties of merchantability or fitness for a particular purpose. It is sold on the understanding that the publisher is not engaged in rendering professional services and neither the publisher nor the author shall be liable for damages arising herefrom. If professional advice or other expert assistance is required, the services of a competent professional should be sought.

Library of Congress Cataloging-in-Publication Data

Bresnan, Joan, author. Lexical-functional syntax / Joan Bresnan. – Second edition. pages cm. – (Blackwell textbooks in linguistics) Includes bibliographical references and index. ISBN 978-1-4051-8781-7 (paperback) 1. Grammar, Comparative and general–Syntax. 2. Lexical-functional grammar. I. Title. P291.B726 2015 415–dc23 2015006866

A catalogue record for this book is available from the British Library.

Cover image: © dpaint / Shutterstock

Set in 10/13pt Sabon by Aptara Inc., New Delhi, India

1 2016 Contents

Preface to the First Edition xi Preface to the Second Edition xv Acknowledgments xvii

I Motivation for the LFG Architecture 1 1 Nonconfigurationality 3 Further reading 10 2 Movement Paradoxes 11 2.1 Theoretical assumptions 15 Further reading and discussion 19 3 Lexicality and Argument Structure 21 3.1 Two approaches to passive relation changes 21 3.2 The lexicality of relation changes 23 3.2.1 English passive verb forms 24 3.2.2 Adjectives versus verbs 24 3.2.3 Participle–adjective conversion 25 3.2.4 Passive participles convert to adjectives 25 3.2.5 Differences between adjectival and verbal passives explained 27 3.2.6 Differences between adjectival and verbal passives unexplained 28 3.2.7 Conclusion: passivization is lexical 32 3.3 Passivization with and without movement 32 Further reading and discussion 36

II Formally Modeling the Architecture 39 4 A Formal Model of Syntactic Structure 41 4.1 Design principles 41 4.1.1 Principle I: variability 41 4.1.2 Principle II: universality 42 4.1.3 Principle III: monotonicity 43 vi Contents

4.2 The definition of f-structures 44 4.3 The description of f-structures 46 4.4 The correspondence between c- and f-structures 48 4.5 The solution algorithm 54 Problems 58 4.6 Defining versus constraining equations 59 4.7 Completeness and coherence 62 Problems 63 4.8 Functional uncertainty 63 4.9 Sets of f-structures 70 4.10 Conclusion 71 Further reading 72 5 Monotonicity and Some of Its Consequences 73 5.1 Monotonicity 73 5.2 Relation changes and monotonicity 76 5.3 Information and form 79 5.3.1 The fragmentability of language 79 5.3.2 The nonconfigurationality of language 82 5.3.3 Apparent information flow through external structure 83 5.3.4 Noncompositionality 84 5.4 Conclusion 85

III Inflectional Morphology and Phrase Structure Variation 87 6 A Theory of Structure–Function Mappings 89 6.1 Grammatical functions 94 6.1.1 Basics of grammatical functions 94 6.1.2 Classification of grammatical functions 100 6.2 The organization of c-structure categories 101 6.2.1 Endocentricity and X′ structures 101 6.2.2 Endocentric mapping to f-structure 104 Problems 111 6.3 Exocentric categories 112 6.3.1 Lexocentricity and S 112 6.3.2 S and endocentricity 115 6.3.3 Nonprojecting words 116 6.3.4 Summary of the structure–function principles 117 6.4 Toward a typology 118 6.5 Effects of economy of expression 119 Further reading and discussion 124 Appendix: X′ theory 125 Contents vii

7 Endocentricity and Heads 129 7.1 Head mobility 129 7.1.1 Verb order in Welsh 130 7.2 Endocentricity and extended heads 135 7.3 Distributed exponence 138 7.3.1 Wambaya c-structure 139 7.3.2 The Wambaya tense system 144 7.4 Conclusion 146 Problems 147 Exercise 147 8 Pronoun Incorporation and Agreement 151 8.1 Chichewaˆ 157 8.1.1 Word order 161 8.1.2 Independent pronouns 162 8.1.3 Contrastive focus 164 8.1.4 Interrogatives and relatives 165 8.1.5 Other syntactic and phonological differences 166 8.1.6 Functional ambiguity of subject and topic 167 8.2 Navajo 171 Exercise 1 180 Exercise 2 180 8.3 Plains Cree and inverse agreement 182 Exercise 3 185 Problems 186 8.4 Two types of agreement: index and concord 186 Exercise 4 192 8.5 Conclusion 192 Further reading and discussion 193 9 and 196 9.1 English topicalization 196 9.2 Russian topicalization 199 9.3 Economy of expression 205 Problems 207 9.4 Topicalization versus scrambling 207 9.5 Detecting empty categories 210 Exercise 223 Further reading and discussion 223 The crossover effect 223 Two types of null pronominals 224 Generalization to operator complexes 225 Other factors 226 viii Contents

IV On Functional Structures: Binding, Predication, and Control 227 10 Basic Binding Theory 229 10.1 Basic concepts 229 10.2 A toy binding theory 231 10.3 Principle C 239 Further reading and discussion 246 10.4 Formalization of the binding constraints 247 11 Types of Bound Anaphors 254 11.1 Dimensions of anaphoric binding 254 11.2 Icelandic: subjective and anti-subjective pronouns 256 11.3 Norwegian: subjective/nuclear pronouns 259 11.4 Logophoricity versus subjectivity 261 Further reading and discussion 273 11.5 The typology of reflexives and the origins of nuclearity 275 Further reading and discussion 283 11.6 Formalization 284 12 Predication Relations 286 12.1 Predicate complements versus adjuncts 286 12.2 F-structures of xcomps 289 Exercise 1 295 Exercise 2 295 12.3 F-structure of PP complements 295 12.4 C-structure of predicate complements 301 12.5 Raising 304 Further reading and discussion 307 13 Anaphoric Control 309 13.1 Gerundive versus participial VPs in English 309 13.2 Structure of gerundive VPs 311 13.3 Anaphoric control versus functional control 319 13.4 Conclusion 323 Problems 323 Further reading and discussion 323 14 From Argument Structure to Functional Structure 324 14.1 What is argument structure? 326 14.2 The theory of a-structures 329 14.3 Mapping a-structures to syntactic functions 333 14.4 Examples and consequences 334 14.4.1 Unaccusatives 334 14.4.2 Resultatives 336 14.4.3 “Fake” reflexives and “nonsubcategorized objects” 336 14.4.4 Word order of internal/external arguments 337 Contents ix

14.4.5 Ditransitives 337 14.4.6 Interactions of passive and raising 340 14.4.7 Morphology that adds or suppresses a-structure roles 341 Problems 344 Further reading and discussion 344

Problem Sets and Solutions 349 Problem Set 1 351 Problem Set 2 354 Problem Set 3 370 Problem Set 4 375 Problem Set 5 391 Problem Set 6 417 Solutions to Selected Problems 436 References for the Problems 461 References 464 Language Index 501 Subject Index 503

Preface to the First Edition

What is LFG? lfg (lexical-functional grammar) is a theory of grammar which has a power- ful, flexible, and mathematically well-defined grammar formalism designed for typologically diverse languages. lfg has provided the framework for a substan- tial amount of descriptive and theoretical research on many languages, including those in Australia (Warlpiri and Wambaya), America (Navajo, Plains Cree, Green- landic), Europe (Welsh, Irish, English, German, Dutch, West Flemish, Icelandic, Norwegian, Finnish, Russian, Serbian/Croatian), Africa (Chichewa,ˆ Ewe, Moroc- can Arabic), South Asia (Malayalam, Hindi), and East Asia (Japanese) which are discussed and analyzed in the text and problem sets of this book.

How is it different? lfg is closely attuned to the overt perceptible expressions of language, and to the abstract relational information that they directly express. lfg has a constraint- based, parallel correspondence architecture; it has no serial derivations (unlike transformational grammar); there are no “deep structures” or “initial structures.” Abstract relations are locally distributed as partial information across words and overt fragments of structure, and may be monotonically synthesized in any order or in parallel. Being designed for a wide range of nonconfigurational and con- figurational language types, lfg departs radically from most other grammar for- malisms in one striking way: it is noncompositional, allowing the “content” of a constituent to vary depending on its context. These descriptions may sound mysterious to the newcomer, but lfg is sim- ple. Field linguists doing primary research on languages have found it easy to use. And because lfg is mathematically well defined and simple, it is also easy to implement. It has been employed in many computational systems, ranging from xii Preface to the First Edition state-of-the-art industrial wide-coverage grammars used for machine translation and processing to pedagogical systems implemented on personal computers. lfg is being used as a representational basis in the new crop of data-driven approaches to language, including Optimality Theoretic syntax and probabilistic analysis of natural language. There is an International Lexical-Functional Gram- mar Association (ILFGA), and there are various websites for lfg resources, the main one being: http://www.essex.ac.uk/linguistics/external/LFG/

What is in this book?

This book provides both an introduction to lfg and a synthesis of major theoreti- cal developments in lexical-functional syntax since the mid to late 1980s. It can be used both as a textbook for students and as a reference text for researchers. Many references to current work are given, but the only background required is some familiarity with elementary formal constructs such as the definition of functions and relations, and an understanding of the basic syntactic concepts of constituent structure and X′ theory (such as in the short paper by Bresnan 1977). The problem sets provide a hands-on way of learning to use the formalism, analytic concepts, and variety of linguistic ideas that can be expressed.

What is not in this book?

Research in lfg is the cooperative effort of an international community of diverse researchers, of which the first author of this book – though one of the original developers of the theory – is just one. The goal of presenting a coherent and acces- sible view of the major developments in lexical-functional syntax has inevitably led to some neglect of important topics and alternatives. The deliberate omissions are these. We have omitted coverage of Optimality Theoretic syntax based on lfg (sometimes called “ot-lfg”); references can be found in Bresnan (2000, 2001a,b) and Sells (2001a), among others, but this area has grown rapidly since the pub- lication of the first edition. We have also omitted any account of probabilistic analysis of language using lfg, such as Data-Oriented Parsing (“dop-lfg”); see Bod and Kaplan (1998), Johnson et al. (1999), and Bod (1999) for several dif- ferent approaches. This book is devoted to lexical-functional syntax and makes no attempt to cover current research in semantics within the lfg framework. (See Dalrymple 1999 for one recent line of research in semantics for lfg.) Constructive Preface to the First Edition xiii

Morphology is also not covered. (See Nordlinger 1998a, Barron 1998, Sadler 1998, Sells 1999, Lee 1999a,b, and Sharma 1999). New developments in coor- dination and feature resolution also had to be omitted because of time and space constraints (Dalrymple and Kaplan 1997, 2000). The history of the development of lfg and its relation to other theories is also omitted and awaits other authors. Despite these omissions, the reader of this book will have no trouble following current research, which can be closely tracked from the website resources men- tioned above.

How to use this book

In teaching lfg from this text, we do not attempt to teach all of the formalism developed in Part II at once. Instead we break Chapter 4 into three parts:

s Sections 4.1–4.5, followed by Problem Set 1; s Sections 4.6–4.7, followed by Problem Set 2; s Sections 4.8–4.9 (read with Chapter 9).

Preface to the Second Edition

The second edition of this book preserves the structure and aims of the first edition, but updates the work by adding expositions of new developments closely related to the topics of the first edition. The latter include new sections on agreement and concord (Chapter 8), on predication (Chapter 12), and on morphology that adds or suppresses argument structure roles (Chapter 14), as well as more explicit anal- yses of the flow of information in conjoined clauses in Wambaya (Chapter 7) and Russian (Chapter 9). There are also additions to the problem sets, including new in-chapter exercises and many problems based on languages not discussed in the first edition. The range of languages that have attracted research inlfg has been steadily expanding and now also includes substantial work on further varieties of Arabic, Biak, Bulgarian, Cantonese and Mandarin Chinese, Dagaare, Danish, Dargwa, Fijian, Ancient and Modern Greek, Hebrew, Hungarian, Indone- sian, Kaqchikel, K’ichee’, Latin, Maltese, Murrinh-Patha, Ossetic, Polish, Brazilian and European Portuguese, Sanskrit, Swedish, Thai, Tigrinya, and Turkish, among many others. The second edition also includes revisions to incorporate the correc- tions and suggestions of many users of the first edition and to remove difficulties for the reader and student. Like the first edition, the second edition does not cover all important topics in lfg. Complex predicates and coordination are two examples of omission. In general, readers of this text will be well equipped to venture into the literature. On the topic of coordination in particular, Peterson (2004) provides a very readable introduction, and the reader interested in details of the analysis can read that paper in conjunction with the discussions of coordinate Ss in Wambaya (Chapter 7) and Russian (Chapter 9). Similarly, like the first edition, the second edition focuses on syntax andsets aside developments in semantics and the syntax–semantics interface in lfg. One major thread of such work concerns glue semantics; see Dalrymple (1999, 2001), Asudeh (2012), and references therein. The first edition preface notes on How to use this book remain as relevant as before. Note that Parts I and II are concerned with the motivation for functional structure, while the theory of lexical and phrasal categories is developed in Part III. Accordingly, the reader should bear in mind that the choice of particular labels xvi Preface to the Second Edition for nodes of phrase structure diagrams plays very little role in Parts I and II. The traditional notions of “clause” and “nominal” are assumed, but whether a clause is labeled S, IP, or CP and whether a nominal is labeled NP or DP have little consequence, until the theory of phrase structure categories is presented in Part III, Chapter 6. There a distinction is drawn between endocentric (headed) IP clauses in languages such as English, versus exocentric (headless) S clauses found in many other languages. Acknowledgments

First edition

Joan Bresnan: It is a pleasure to thank the friends and colleagues who gave me detailed comments on earlier drafts of individual chapters or even the entire manuscript of this work: Farrell Ackerman, Avery Andrews, Judith Berman, Chris Culy, Mary Dalrymple, Yehuda Falk, Paul Kroeger, Helge Lødrup, Elaine Malkin and Stella Markantonatou, Chris Manning, Michiko Nakano, Louisa Sadler, and Peter Sells. The original problem sets and solutions were first edited in 1987 by Kristin Hanson, with the assistance of Heinrich Beck, Nick Clements, Mary Dalrymple, Jeffrey Goldberg, Jonni Kanerva, Yo Matsumoto, James McCloskey, Mariko Saiki, Stephen Wechsler, Michael Wescoat, Annie Zaenen, and Draga Zec. Subsequent contributions were made by Rachel Nordlinger and Ida Toivonen, and Ash Asudeh converted the files toA LTEX. I owe special thanks to Ron Kaplan for answering questions about formal issues with his usual blend of generosity and rigor, and in providing through Xerox PARC some financial assistance in the research and preparation of the manuscript. For the personal support which made it possible for me to finish writing this book, and for the inspiration of living with a finisher and a champion, I also thank Marianne.

Second edition

All: We thank the following readers of the first edition for specific comments and corrections, which we have tried to address: Judith Berman, George Aaron Broad- well, William Blunk-Fernandez, Mary Dalrymple, Yehuda Falk, Itamar Francez, One-Soon Her, Florian Jaeger, Ron Kaplan, Tracy Holloway King, Paul Kroeger, Tibor Laczko,´ Joan Maling, and Peter Sells. We thank Adam Przepiorkowski´ for comments and corrections on a draft of the second edition. We also thank xviii Acknowledgments the anonymous reviewers of the second edition and our patient editors at Wiley, Danielle Descoteaux and Julia Kirk. Any remaining errors are our own. We thank students at Carleton University, Oxford University, Stanford Univer- sity, and UT Austin for their feedback and questions on drafts of the second edition used during teaching. We thank Julian Michael from UT Austin, in particular, for a comprehensive list of notes and errata. Crystal Bruce, Liz Christie, and Paul Melchin supported preparation of the sec- ond edition through research assistantships at Carleton University. We thank Paul in particular for carefully reading drafts, including flagging errors and inconsis- tencies and raising many interesting questions. Once again, any remaining errors are our own. For their help with problem sets and data, we would like to thank Alex Alsina and Sam Mchombo (Chichewaˆ problem), Telma Can Pixabaj and Nora England (K’ichee’ problem), and Samu Niskanen (Finnish data).

Ash Asudeh: I acknowledge the support of three grants/awards which contributed to the preparation of this volume: an Early Researcher Award from the Province of Ontario, a Standard Research Grant from the Social Sciences and Humani- ties Research Council of Canada (#410-2006-1650), and an Individual Discovery Grant from the Natural Sciences and Engineering Research Council of Canada (#371969). For personal support, I would like to thank Alli Asudeh, Isa Asudeh, Behin Ferguson, George Ferguson, Esra Kaytaz, Shahnaz Khorshidi, Elliott Strikefoot, and Ida Toivonen.Last but most, I thank our kids, Thora, Isak, Alfred, and Harald, for everything. My efforts on this volume are dedicated to them.

Ida Toivonen: I acknowledge the support of a Standard Research Grant from the Social Sciences and Humanities Research Council of Canada (#410-2010-1841). I want to thank my family and friends for all their help and support.

Stephen Wechsler: The writing of this book was supported in part by summer funding from the Alma Madden Cowell Professorship in Linguistics at the Uni- versity of Texas. I am grateful for that support. Part I Motivation for the LFG Architecture

Introduction

The search for a universal design of grammar has long motivated research in lin- guistic theory. Language is both universal among humans and specific to us. Any child can acquire fluent mastery of any of the thousands of human languages, given sufficient exposure, but no animal has this capacity. These simple factshave suggested to many linguists that there must be a universal design of grammar, a common organizing structure of all languages that underlies their superficial vari- ations in modes of expression. If this universal grammar is a biologically given form of knowledge, as many linguists assume today, then study of the invariants of the structure of human languages may tell us something fundamental about the human mind. This rationalist, universalist conception of linguistics has a long intellectual tra- dition, appearing in the works of philosophers and grammarians of the past six centuries. In the past century it has been revived by Noam Chomsky. Chomsky’s great achievement is to couple the universalist conception of language from the tradition of philosophical grammar with a far more precise model of linguistic structure adapted from the mathematics of formal systems developed in the twen- tieth century. This powerful combination of ideas, called “generative grammar,” has revolutionized linguistic theory. In the methodological paradigm of genera- tive grammar, formal representations of linguistic structures are developed and empirically tested against native speakers’ knowledge of their language. Universal grammar limits the space of formal structures. Generative grammar holds that language cannot be adequately characterized solely in terms of a formal description of its overt constituents, or “surface

Lexical-Functional Syntax, Second Edition. Joan Bresnan, Ash Asudeh, Ida Toivonen, and Stephen Wechsler. © 2016 Joan Bresnan, Ash Asudeh, Ida Toivonen, and Stephen Wechsler. Published 2016 by John Wiley & Sons Ltd. 2 Motivation for the LFG Architecture structure.” A more abstract representation is also needed to represent the implicit linguistic knowledge of speakers. Chomsky has conceived of this abstract repre- sentation as a “deep” or initial structure which undergoes sequential serial oper- ations (transformations) to derive the overt perceptible form. It is to explain how these abstract formal structures are acquired by speakers that Chomsky developed his rationalist epistemology: human beings possess an innate faculty specialized for language which enables them to acquire complex human languages despite the poverty of stimulus in their learning environment.1 Towards the end of the twentieth century, new formal ideas began to achieve prominence in linguistic theory, making use of parallel rather than serial structures and computations, and comparative evaluation of multiple overt structures rather than serial derivation of individual overt structures. These ideas are embodied in a family of nonderivational, constraint-based linguistic theories and in approaches based on optimization (both statistical and discrete). These newer theories are compatible with different linguistic epistemologies drawing on structuralist and functional/typological ideas which have both predated and coexisted with gener- ative grammar. One such theory is lexical-functional grammar (lfg) (Kaplan and Bresnan 1982), which is the topic of this book. Part I of this book empirically and informally motivates the lfg architecture by looking at the core linguistic phenomena which inspired it: nonconfigurationality, movement paradoxes, and the lexicality of relation changes such as passivization. Part II shows how the intuitive ideas of Part I can be formally modeled as flexible correspondence mappings between parallel structures (constituent structure and functional structure). Part III presents a theory and typology of structure–function correspondences, together with several case studies of languages in which syntac- tic functions are created morphologically rather than by constituent structures. Part IV motivates functional structure by showing how invariances of language are captured in functional structures and outlines a theory of how functional struc- tures are projected from argument structures.

1 Chomsky’s “poverty of stimulus” argument for universal grammar has attracted criticism (e.g. Van Valin 1994, Pullum 1997, Pullum and Scholz 2002). The most controversial aspect is not the conclu- sion that humans have innate biological capacities or learning biases that support language – no one doubts that the phonological structure of language is supported by our innate articulatory and per- ceptual systems – but the assumption that such capacities must be narrowly specialized for acquiring grammatical systems – and grammatical systems of the specific types advocated by Chomsky. It is also true that advances in statistical learning theories permit inferences about nonoccurring data which enrich the informativeness of the stimulus (e.g. Tesar and Smolensky 1998, Boersma and Hayes 2001; for discussion of empirical case studies, see Lidz et al. 2003, Foraker et al. 2009, Reali and Christiansen 2005, Ambridge et al. 2008, and Kam et al. 2008). 1 Nonconfigurationality

One fundamental problem for the design of universal grammar is the great vari- ability in modes of expression of languages. Languages differ radically in the ways in which they form similar ideas into words and phrases. The idea of two small children chasing a dog is expressed in English by means of a phrase structure in which conceptual components of the whole – the concept of the two small chil- dren and the concept of the dog being two such components – correspond to single phrases. Phrases are groups of contiguous words that are units for substi- tutions, remain together as units under stylistic permutations and paraphrases of a sentence, constrain the pronunciation patterns of sentences, and are subject to ordering constraints relative to other words and word groups. The (simplified) phrase structure of an English sentence is illustrated in (1):1

(1) S

NP VP

the two small children Aux VP

are V NP

chasing that dog

1 For simplicity, the root node in (1) bears the traditional label S (“sentence”), instead of a symbol for a phrasal projection, such as IP (“inflection phrase”). For English, S will be replaced with IP later.More generally, in Parts I and II of this book (Chapters 1–5), the particular categories labeling the nodes of phrase structure diagrams play very little role. The traditional notions of “clause” and “nominal” are assumed, but whether a clause is labeled S, IP, or CP and whether a nominal is labeled NP or DP have little consequence, until the theory of phrase structure categories is presented in Part III, Chapter 6. There a distinction is drawn between endocentric (headed) IP clauses in languages such as English and exocentric (headless) S clauses found in many other languages.

Lexical-Functional Syntax, Second Edition. Joan Bresnan, Ash Asudeh, Ida Toivonen, and Stephen Wechsler. © 2016 Joan Bresnan, Ash Asudeh, Ida Toivonen, and Stephen Wechsler. Published 2016 by John Wiley & Sons Ltd. 4 Nonconfigurationality

In this structure, the word combinations the two small children and that dog are noun phrases (NPs), in which the words cannot be separated, and there is also a verb phrase (VP). When the phrases are freely broken up, the result is ungram- matical or different in meaning:

(2) a. ∗The two small are chasing that children dog. b. ∗The two small are dog chasing children that. c. ∗Chasing are the two small that dog children. d. ∗That are children chasing the two small dog.

The simple correspondence between conceptual units and grammatical phrases seems so natural to the English speaker as to appear a necessary feature of lan- guage itself – but it is not. Consider Warlpiri, a language of the people who have inhabited Australia since long before the colonization of that continent by English speakers.2 Example (3) shows the phrase structure of a Warlpiri sentence express- ing the same idea as the English sentence (1).3 But in Warlpiri, every permutation of the words in the sentence is possible, with the same meaning, so long as the auxiliary (Aux) tense marker occurs in the second position. In particular, the word orders of all the bad English examples in (2) are good in Warlpiri.

(3) S

NP Aux V NP NP NP

wita-jarra-rlu ka-pala wajilipi-nyi yalumpu kurdu-jarra-rlu maliki small-DUAL-ERG pres-3DU.SUBJ chase-NPAST that.ABS child-DUAL-ERG dog.ABS

It is not true that Warlpiri lacks phrases altogether: syntactic analysis has shown that some phrases (NPs but not VPs) do optionally occur, and there is evidence for a somewhat more articulated clause structure including a focus position to the left of Aux.4 What appears to the left of Aux may be a single word, as in (3), or a single multi-word NP, which then allows the nonfinal case-marker to be omitted, as in (4):

(4) Kurdu-jarra(-rlu) wita-jarra-rlu ka-pala maliki wajilipi-nyi. child-dual(-erg) small-dual-erg pres-3du.subj dog.abs chase-nonpast “The two small children are chasing the dog.”

2 See Dixon (1981) on the history and nature of Australian languages. Of the English colonization of Australia he writes, “I have chosen to write plainly, to talk of the white ‘invasion’ of Aboriginal Australian lands, avoiding euphemisms such as ‘settlement’, and not to gloss over murder – of people, tribes, and languages”(Dixon 1981, p. xiv). The Warlpiri grammatical phenomena discussed here come from Hale (1981) and Simpson (1983a, 1991). 3 Note that the tree structure in (3) is oversimplified for expository purposes. Compare it with the phrase structure typology in Chapter 6 and the analysis of Wambaya in Chapter 7. 4 See Austin and Bresnan (1996) and Simpson (2007) for details. Nonconfigurationality 5

But crucially, the sole item appearing to the left of Aux cannot be a VP,nor is there any other evidence that VP is ever a phrasal constituent in Warlpiri.5 The subject of a Warlpiri sentence is not identified by its position in the phrase structure, as it is in English, but rather by the appearance of the ergative case marker -rlu. More generally, phrases are not essential to the expression of conceptual units. The coherence of a conceptual unit in Warlpiri is indicated by means of word shapes rather than word groups: noncontiguous words that form a conceptual unit must share the same formal endings – case and number morphology. In (3) the word for ‘small’ shares the dual and ergative endings -jarra and -rlu with the word for ‘child’ which it modifies, and these endings differ from those of the words for ‘dog’ and ‘that’, which are null. Thus the words kurdu-jarra-rlu (‘child-dual- erg’) and wita-jarra-rlu (‘small-dual-erg’) jointly express the concept of ‘two small children’ and jointly serve as subject of the sentence – regardless of whether those words appear together as a constituent (as in (4)) or not (as in (3)). This difference between Warlpiri and English exemplifies a broad crosslinguis- tic generalization observed by many students of linguistic typology: across lan- guages, there often appears to be an inverse relation between the amount of gram- matical information expressed by word structure and the amount expressed by phrase structure. Words must appear in a sequence since they cannot be pro- nounced simultaneously, so the relative order of words is generally available as one means of expressing the grammatical relationships necessary for communi- cation. But some languages lack the rich word structure found in languages such as Warlpiri. Thus languages rich in word structure (morphology) may make more or less use of fixed phrase structure forms (syntax), whereas languages poorin morphology overwhelmingly tend to have more rigid, hierarchical phrase struc- tures. This trade-off between morphology and rigid phrase structure is spec- tacularly illustrated by some of the radically nonconfigurational languages of Australia, but there is evidence for it also in the other language types we will examine in Part III. We can summarize this generalization with the slogan “Mor- phology competes with syntax”for the job of expressing the grammatical relations between words. The idea that words and phrases are alternative means of expressing the same grammatical relations underlies the design of lfg and distinguishes it from other formal syntactic frameworks. In addition, we cannot discount the effect of “con- figurational bias.” Through historical accident, the resources of modern science and technology have been dominated by states whose national languages happen to be highly configurational. As a result, there has been a vast lack of knowledge of

5 The Aux in Warlpiri generally follows the first phrase of the sentence; the parts of a noun phrase can appear together in this position, but not the parts of a verb phrase. See Simpson (1991, 2007), Austin and Bresnan (1996), and Nordlinger (1998a) for more detailed discussion. With special articulation of discourse information structure, multiple noun phrases may appear, but never a verb phrase (Legate 2002; Laughren 2002). These facts are consistent with the constituent structure theory developed in Chapter 6 and in Simpson (2007). 6 Nonconfigurationality typological variation of language within the scientific establishment in computer science, logic, and philosophy – and even among many theoretical linguists of a formal bent. Although Warlpiri lacks English-style phrase structure, and English lacks Warlpiri-style case and agreement forms of words, there is evidence that they have a common organization at a deeper level than is apparent from their dif- fering modes of expression. Similar conceptual units are expressed by the two languages – objects and their relations and motions, events and their participants, and human emotions, actions, and aims. And at an appropriate level of abstrac- tion, similar grammatical constraints emerge.6 For example, in English, a reflexive pronoun can be an object coreferring with the subject, but cannot be a subject coreferring with the object:

(5) a. Lucy is hitting herself. b. ∗HerselfishittingLucy.

ThesameistrueinWarlpiri:

(6) a. Napaljarri-rli ka-nyanu paka-rni. Napaljarri-erg pres-refl hit-nonpast “Napaljarri is hitting hereself.” b. ∗Napaljarri ka-nyanu paka-rni. Napaljarri.abs pres-refl hit-nonpast “Herself is hitting Napaljarri.”

This constraint holds in Warlpiri whether or not the subject is discontinuous. Indeed, this grammatical constraint on reflexive pronouns is shared by many lan- guages (see Chapters 10 and 11 and references). Thus, while phrase structure does not universally correspond to conceptual structure, the more abstract grammatical functions it expresses – such as subject and object – are widely shared by different languages. These grammatical func- tions represent classes of varying forms of expression that are equivalent under the correspondence mappings to argument structure (discussed below). Here is the first choice point in the design of universal grammar: how tocap- ture the abstraction of grammatical functions, such as subject and object, across the rather different means of expressing them? The overwhelmingly predominant tendency in modern linguistic theory – due to Chomsky – has been to define them as the familiar configurations of English phrase structure: the subject is anNPin configuration (7a), and the object is an NP in configuration (7b):

6 The following illustration is from Simpson (1983a); see also Hale (1973). Nonconfigurationality 7

(7) a. S

NP ...

b. VP

V NP

For a language like Warlpiri, this choice amounts to the claim that it does have English-style phrase structure after all – not on the surface, where conditions on word order hold, but at an underlying level of representation at which the gram- matical conditions on reflexive pronouns hold.7 Let us refer to this as the config- urational design of universal grammar. It is illustrated in (8)–(9):

(8) English: ⇒ S S NP VP NP VP Aux VP Aux VP V NP V NP

(9) Warlpiri: ⇒ S S NP VP NP Aux V NP NP NP Aux VP V NP

Under the assumption of the configurational design of universal grammar, a reflexive pronoun must satisfy the following grammatical condition onthe underlying level of structure: the reflexive must be contained in a constituent that

7 This hypothesis has taken several forms. One is that the underlying structure is a deep structure, which undergoes transformational “scrambling” rules to derive the modes of expression peculiar to Warlpiri (Hale 1973). But as Hale (1994) points out, the complete absence of movement rules elsewhere in Warlpiri renders this hypothesis unattractive. Another is that the underlying structure is a “lexical structure” which is not transformationally related to the surface forms but represents the universal component of Warlpiri grammar in the phrasal forms of English (Hale 1983). This view has been criticized by Speas (1990) for adopting different theories of grammar for different language types, and has subsequently been abandoned by Hale (1994). A third and more recent form of the hypothesis, based on work by Jelinek (1984) and Baker (1991), assumes that overt NPs are not arguments but adjuncts to incorporated pronouns; see Austin and Bresnan (1996), Nordlinger (1998a), and Croft (1997) for criticism of the latter hypothesis. 8 Nonconfigurationality contains the verb also but not the antecedent of that pronoun; that constituent is the VP. Using the graphic representation of the tree diagram, this means that the antecedent must be “higher” in the tree than the pronoun that it binds. Hence a subject can bind an object but an object cannot bind a subject, whether in English or Warlpiri, as we saw in (5) and (6). This is one approach to capturing the simi- larity between the two languages – but is it the right approach? Now it might be true that all languages do have an abstract level of grammat- ical structure which closely matches the surface organization of the grammars of English and other European languages. (Perhaps it just happens that the biologi- cally based universal design of grammar really does have the form of the language of the colonizers.) But there is no evidence of this; for example, none of the prop- erties of phrases that we mentioned – contiguity under permutation, grouping for pronunciation, ordering relative to other elements, and substitutability – supports the existence of a VP in Warlpiri, and what evidence there is for phrases in Warlpiri shows clearly that there is no VP in our original sense (as discussed in the references previously mentioned: Simpson 1983a, 1991, 2007, Austin and Bresnan 1996, and Nordlinger 1998a). Moreover, there is evidence that the constraints on reflexive pronouns depend not directly on phrase structure configurations but on factors such as predication relations, which are at best only partially reflected in phrase structure configurations (see Kroeger 1993, Manning 1996, Wechsler and Arka 1998, and Part IV of this book). Therefore the “deep” or underlying VP that must be postulated in (9) is devoid of the original constituency properties of VPs. Hence, an alternative taken in the development of lfg is to choose a more abstract representation of the grammatical functions subject and object, one which is neutral between the differing modes of expression of languages. On this alter- native, grammatical functions are not reducible to phrase structure configurations as in (7); they are classes of differing formal expressions that are mapped into argument structure in equivalent ways. Thus we have a differing picture of the grammatical structures of English and Warlpiri:

(10) English: chase 〈 agent patient 〉

PRED SUBJ OBJ

S

NP VP

... Aux VP

V NP

... Nonconfigurationality 9

(11) Warlpiri:

chase 〈 agent patient 〉

PRED SUBJ OBJ

S

NP Aux V NP NP NP

...-ERG ...ABS ...-ERG ...ABS

In this design, the grammatical functions subject and object are equivalence classes that serve as the relators, or “links,”between two formally different, parallel struc- tures: (i) the argument structure, which includes the participants in events and situations, and the relations between those participants, that are grammatically expressed; and (ii) the expression structure, which consists of the modes of expres- sion of the language, represented as a constituent structure (tree). Returning to our example of the reflexive pronouns, the grammatical conditions are now stated directly in terms of those grammatical functions: whether in English or Warlpiri, the subject (subj) can bind the reflexive object (obj) but not vice versa. The lan- guages differ, however, in the way those grammatical functions are expressed. While phrase structure configurations distinguish the subject and object functions in English, the case inflections – erg(ative) and abs(olutive) – distinguish the same functions in Warlpiri. These functions differ overtly, as we have seen, but they show a similar system of correspondences to the argument structure. The system of functions that relates these two structures has been mathematically modeled by the functional structures of lfg (Chapter 4). Let us refer to this as the relational design of universal grammar. Does this choice of representations for grammatical functions make any differ- ence, or are they just notational variants? In fact, there are interesting empirical consequences of the choice of design. The configurational design implies that spe- cific elements of phrase structure – NPs, VPs, and their relations – appearnot only in representing the modes of expression of English and similar European languages, but also in representing deeper aspects of grammatical organization – the abstract syntactic functions and the semantic predicate argument structures of all languages. The relational design, in contrast, implies that it is the distinctive structure of predicators and arguments and their grammatical functions that are relevant at the deeper levels. In the course of this book, we will see many cases where these two designs lead to divergent expectations about the grammars of human languages. In conclusion, an important source of empirical motivation for the relational design of universal grammar adopted by lfg is the existence of phrase structure nonconfigurationality. Although various degrees of nonconfigurationality occur 10 Nonconfigurationality across languages, as we will see in Part III, a number of the Australian lan- guages are among the best exemplars of the phenomenon (Simpson 1991, 2007; Austin and Bresnan 1996; Nordlinger 1998a). This nonconfigurationality is pos- sible because the same grammatical information can be specified by word shapes as by word groups; the functional structure of lfg characterizes this grammat- ical information in an abstract, neutral way, without configurational bias. Thus in “lexical-functional grammar” the term “lexical” refers to the fundamental fact that words, or lexical elements, are as important as syntactic elements in express- ing grammatical information, and the term “functional” refers to the fact that this grammatical information is not identified with particular structural forms of expression, but is viewed as a system of abstract relators of expressions to eventualities.

Further reading

For further readings that examine the evidence and issues of nonconfigurational- ity in more detail, Austin and Bresnan (1996), Nordlinger (1998a), and Simpson (2007) are recommended as particularly accessible. 2 Movement Paradoxes

If typological considerations such as nonconfigurationality motivate a relational design of universal grammar, it is natural to ask how the evidence for syntactic transformations that we see in English and related languages can be accounted for within the same design of grammar. In these familiar languages, we see overt evidence that phrase structure constituents can be displaced from one position to another. The evidence takes the form of extraction configurations. An extraction configuration is illustrated by the set of sentences in (1):

(1) a. We talked about that problem for days. b. That problem, we talked about for days. c. ∗We talked about for days.

In an extraction configuration, a displaced constituent is identified with a lexically empty position. Thus in (1b) the initial NP that problem appears to have been displaced from the usual position of a prepositional object, shown in (1a), to its sentence-initial position. The sign of displacement is the presence of a lexically empty position or gap in the clausal residue, which would be ill-formed on its own, as shown in (1c). An extraction configuration motivates the idea that the sentence-initial constituent in (1b) is moved from an underlying position where it is required:

(2) [That problem] we talked about for days.

Such extraction configurations seem to be direct evidence of the existence of syntactic transformations – structure-dependent operations that map underlying phrase structures representing grammatical relations onto the transformed struc- tures that represent the surface forms of expression of language. If such move- ments are needed in English, it has been argued, then theoretical considerations

Lexical-Functional Syntax, Second Edition. Joan Bresnan, Ash Asudeh, Ida Toivonen, and Stephen Wechsler. © 2016 Joan Bresnan, Ash Asudeh, Ida Toivonen, and Stephen Wechsler. Published 2016 by John Wiley & Sons Ltd. 12 Movement Paradoxes of simplicity, universality, and learnability in principle of language require them in all languages. This conclusion would support the configurational design of grammar. In fact, we will see below that such apparent movements arise in the lfg archi- tecture from the interaction of general principles of structure–function correspon- dence, and can be explained without any loss of generality.Even more interestingly, it turns out that certain “movement paradoxes” favor this architecture over the movement transformations of the configurational design. We can conclude that the same theoretical considerations of simplicity, universality, and learnability in principle require the relational design of grammar for all languages. A movement paradox is illustrated in (3). Note that there is a mismatch in category type between a seemingly moved phrase, the initial that complement in (3a), and the position it is supposed to be moved from, as illustrated in (3b). In (3) the moved phrase cannot appear in the position from which it is sup- posed to be moved, because it is of the wrong type of syntactic category for that position:

(3) a. [That he was SICK]S we talked about for days.

b. *We talked about [that he was sick]S for days. c. cf.We talked about [the fact that he was sick]NP for days.

The complex noun phrase in (3c) is fine as a complement to about. This preposition requires an NP complement, which is why (3b) is not grammatical (though it is perfectly interpretable).1 A related example of a movement paradox involves passivization of sentential complements. It is commonly assumed that the subject of a passive verb is moved from an underlying object position, as in (4):

(4) a. This theory captures that fact. b.[That fact] is captured by this theory.

But with sentential complements, this assumption breaks down with certain verbs for many speakers. The passive example (5a) is a grammatically well-formed sen- tence with a sentential subject, but the corresponding active (5b) is ill-formed for

1 English differs in this respect from some other Germanic languages, such as Swedish. See Berman (1997) for a different analysis of similar facts in German within lfg. Note also that even in English, interrogative clauses are more nominal in character, appearing as prepositional complements: We talked about whether you should be the one to go. Also, in some dialects of English examples like (3b) do occur. Movement Paradoxes 13 many speakers for whom the verb capture (like the preposition about) requires an NP object, as in (5c):

(5) a. [That languages are learnable]S is captured by this theory.

b. *This theory captures [that languages are learnable] S . c. cf.This theory captures [the fact that languages are learnable]NP.

How then could the sentential subject in (5a) have been moved from the object position in (5b)? Another movement paradox occurs with VP preposing. In VP preposing, a topi- cal VP complement to an auxiliary is shifted to the front of the sentence to empha- size the auxiliary, as in (6b). The specific verb form of the fronted phrase depends on the requirements of the auxiliary, as shown by the contrasts between (6a–c):

(6) CONTEXT: I said I would meet you, ...

a. ...and [meet you] I WILL !

b. *... and [met you] I WILL ! c. *... and [ meeting you] I WILL !

A movement approach requires that the moved constituent be identical to the consituent in its original position. The grammaticality or ungrammaticality of (6a–c) follows directly from that of the corresponding sources of movement shown in (7a–c), respectively:

(7) a. I will [meet you]. b. ∗I will [met you]. c. ∗I will [meeting you].

With the perfect auxiliary, however, there can be a mismatch between the fronted verb form (8a) and the in situ verb form (8b) (Andrews 1994, p. 218):

(8) a. She said she would meet me, and [meet me] she HAS !

b. *...and she HAS [meet me ]! c. ...and she HAS [met me]!

How then could the VP in (8a) be moved from a postauxiliary position where it can never occur, as in (8b)? The perfect auxiliary have in English requires that its complement verb be a past participle, as the contrast between (8b) and (8c) illustrates. Paradoxically, then, an infinitive VP cannot in fact appear in the source 14 Movement Paradoxes position from which it is assumed to have been moved. Nevertheless, it is quite natural in English for the infinitive form of the verb to appear in the fronted posi- tion, creating another movement paradox. Subject–auxiliary provides yet another kind of movement paradox. The contracted negative present-tense auxiliary be in Standard English has a gap in its person–number paradigm, shown in (9):

(9) sg. pl. 1 aren’t 2 aren’t aren’t 3 isn’t aren’t

A form exists for every combination of person and number except for first-person singular. Although nonstandard dialects have forms such as ain’t or amn’t which fill this gap, the Standard dialect does not make use of these forms:

(10) a. I am not your friend. b. ∗I ain’t/amn’t your friend.

Under subject–auxiliary inversion, many speakers of the Standard dialect find that it is natural to use the form aren’t (Langendoen 1970; Hudson 1977; Dixon 1982; Gazdar et al. 1982; Kim and Sag 1996; Bresnan 2000, 2001b; Bresnan et al. 2007):

(11) a.Am I not your friend? (stilted or very formal)

b. Aren’t I your friend?

If the inverted auxiliary in (11a) is derived by movement from an underlying post- subject position, where is it moved from? The form aren’t cannot appear in the postsubject position:

(12) ∗I aren’t your friend.

Movement paradoxes are problematic for two basic architectural assumptions of transformational theories: the representational assumption that the inner or underlying structure of language has the formal categorial properties of phrase structure (the configurational design of grammar), and the derivational assump- tion that the surface configurations of phrase structure categories are derived by transformational operations (movements) from basic syntactic representations of the same type.2

2 For some of these problems, various proposals have been put forward within the movement frame- works, but none of them is completely satisfactory and there is little agreement on the solutions. See the Movement Paradoxes 15

2.1 Theoretical assumptions

The kinds of category mismatches we see in the movement paradoxes are expected in the relational architecture of lfg, for the simple reason that the correspondence between structure and function is not perfect: there can be mis- matches between the functional structure attributes of an element and the constituent structure positions it can appear in. Functional structure (f-structure) is a feature structure that represents, among other things, grammatical functions. Constituent, or categorial, structure (c-structure) is a tree structure that contains categorial information; see Chapter 4 for more on f-structure and c-structure. We consider here just one example, in order to illustrate the principle. The diagram in (13) again informally depicts the architecture illustrated in (10) and (11) of Chapter 1:

(13) Parallel structures (LFG):

argument (a-)structure: verb 〈 x , y 〉

⎡ ⎤ ⎢ . ⎥ ⎢ SUBJ . ⎥ ⎢ ⎥ functional (f-)structure: ⎢ . ⎥ ⎣ OBJ . ⎦ PRED ...

V

constituent (c-)structure: V NP

N

Each structure models a different dimension of grammatical substance: role, function, or category. Roles correspond to the grammatically expressible partic- ipants of eventualities (modeled by a-structure), syntactic functions belong to the abstract system of relators of roles to expressions (modeled by f-structure), and phrase structure categories belong to the overt structure of forms of expression (modeled by c-structure). The structures are associated by principles of functional correspondence (also called “linking” or “mapping” principles; see Chapters 6 and 14).

“Further reading and discussion” section for references. The movement paradoxes with VP preposing and auxiliary inversion have generally been ignored within the movement approaches. The former is analyzed by Andrews (1994) within lfg and the latter by Bresnan (2000, 2001b). 16 Movement Paradoxes

Let us take example (5a) by way of illustration.3 The surface constituent struc- ture can be depicted as in (14):4

(14) IP

CP IP

that languages are learnable I

IVP

is V PP

captured by this theory

Observe that the that complement is base-generated in topicalized position as a sentential complement (CP) adjoined to the sentence (IP) in the c-structure. The evidence for a subject-external position of the CP in English includes the observa- tion that auxiliary inversion with the CP is not possible (Koster 1978):

(15) a. That he’ll be late is quite likely. b. ∗Is that he’ll be late likely? c. ∗How likely is that he’ll be late?

General principles of structure–function correspondence for configurational struc- tures (to be discussed in Chapter 6) define the position of the adjoined CP as that of a topic (top), focus (foc), or adjunct (adj) function, and tell us that the entire sentence corresponds to a single functional structure nucleus.5 Because the CP is a non-nominal category, it cannot occupy the subject NP position as sister of I′. This follows from the constraint on structure–function association shown in (16) (Bresnan 1994b, ex. 88):

(16) The canonical structural realization of subject and object functions is nominal.

Although the CP cannot appear in the usual structural position of the subject, which is an NP position by (16), general principles of the theory nevertheless link the function of the topicalized phrase with the subject function in f-structure,

3 Grimshaw (1982a) sketches the essential outlines of the solution adopted here. 4 We use the X′-theoretic functional category I and its projections, I′ andIP.IPcanbereadasSandI as Aux, if this is preferred. 5 – corresponding to what Grimshaw (1991) calls an “extended projection.” Movement Paradoxes 17 as shown in (17). (The phrases in double quotation marks abbreviate the actual f-structure contents where they are not relevant to the discussion.) The first princi- ple is simply that the subject is the default topic of its sentence, which we formalize (in Section 6.2.1) by optionally identifying subj with top:6

(17) Correspondences between c- and f-structures of (5a): ⎡ ⎤ TOP “that ...” ⎢ ⎥ ⎢ SUBJ ⎥ ⎣ ⎦ IP PRED ‘capture SUBJ, OBLθ ’ OBLθ “by this theory” CP IP

that ... is VP

V PP

captured by this theory

As we will see in Chapter 4, the general principle of completeness requires that the verbal a-structure requirements, including the demand for a subject, be met; so something must fill the subj function in f-structure. Coherence requires that the grammaticalized discourse functions (top) and (foc) be identified with a syntactic function such as subj or obj; so the topicalized CP must fill some other function in the clause. Uniqueness prevents identifying the top with a syntactic function borne by a lexically filled constituent in c-structure, as we will see in Chapter 4; sothere must be an unfilled function – a gap – for the topic to fill. The gap is represented as the omission of the subject NP from its normal c-structure position as sister of I′. In this way the typical elements of an extraction configuration – namely, the identification of a displaced constituent with a lexically empty position –arise nonderivationally from general principles. Finally, the passive relation follows not from NP movement but from princi- ples of a-structure to f-structure correspondence provided by the lexical mapping theory. This part of the theory will be discussed in later chapters, but we briefly illustrate the main point in (18). The passive a-structure is morpholexically asso- ciated with the passive participle of capture; it differs from the active in that the agent role (the x argument in diagram (13)) is suppressed, and may be associated 7 with an optional by-phrase (obl𝜃). The a-structure allows the patient role (the y argument in (13)) to be associated with either the subj or the obj function,

6 The formal mechanism of functional identification of displaced constituents with other functions is discussed in Chapters 4, 6, and 9. 7 It may be preferable to instead analyze the by-phrase as an adjunct, but that would not substantively affect the point at hand; see Needham and Toivonen (2011) for further discussion. 18 Movement Paradoxes depending on context. It is the suppression of the agent that allows the theme role to correspond to the subj function in this case:

(18) Morpholexical correspondences between a- and f-structures of (5a):

a-structure: capture 〈 x y 〉

⎡ ⎤ PRED ⎢ ⎥ ⎢ . ⎥ f-structure: ⎣ SUBJ . ⎦ . OBLθ .

Thus, the reason there can be a mismatch between the topicalized category CP and the object position of the verb capture is that the topicalized that complement is flexibly mapped to its argument role by the correspondence theory; itisnot linked to its argument role by a series of transformational movements from an underlying object position in phrase structure. And the reason for this is that the logical roles of phrases are not uniformly represented in phrase structure form. Role, function, and category are independent dimensions of linguistic substance. Within the constraints on their correspondences, mismatches are possible. Why then is the active form (5b) ∗This theory captures that languages are learn- able ill-formed? The answer is simply that capture takes only objects as its com- plements and by (16) the canonical structural realization of objects is a nominal category, not a that complement CP. When English verbs take that complements, these complements are not objects, and the verbs are syntactically intransitive in this construction. The verb care is an example of the latter type:8

(19) a. I don’t care that languages are learnable. b. ∗I don’t care that. c. ∗That languages are learnable isn’t cared.

Observe that this architecture leads us to expect nonuniformities of syntactic categorization of parallel semantic categories across languages: if English senten- tial complement CPs can have the grammatical functions of subject and object in f-structure, as proposed here, and if subject and object are canonically realized as NPs in c-structure, as we have also hypothesized, then we should expect to find languages in which there are sentential complement NPs in c-structure. In fact, Bresnan (1995a) argues that Chichewaˆ has nominal kutiˇ (‘that’) complements,

8 Some verbs, such as believe, share properties of both capture and care.Likecapture, believe takes an object which alternates with a passive subject. But like care, believe allows a nonobject sentential complement. See Grimshaw (1982a) and Bresnan (1995a) for further discussion of this analysis. Movement Paradoxes 19 and there is evidence that this is true also for Dutch (Hoekstra 1984) and Spanish (Plann 1986; Lujan´ 1994), to give just a few relevant instances.9 In conclusion, we have seen in principle how both the typologically “exotic” property of nonconfigurationality and the category mismatches in extraction con- figurations found in English are comprehended by the same relational architecture of universal grammar. Chapters 4 and 5 present an explicit model of parallel struc- tures and their correspondence mappings to make these linguistic ideas precise. The next chapter discusses our third motivation for the relational architecture, based on the lexicality of passivization and other changes in argument structure realization.

Further reading and discussion

Further readings that discuss the issue of category mismatches and imperfect cor- respondence between structure and function in a particularly accessible way are Bresnan (1994b, 1995a) on English, Mohanan (1994, 1995) and Butt (1995) on Hindi/Urdu, Matsumoto (1996a) on Japanese, and Broadwell (2010) on Zapotec. Quite intricate argumentation is involved in evaluating the various movement approaches. For that-clause topicalization, analyses have been proposed which involve: (i) movement and deletion of an underlying resumptive pronoun (e.g. Koster 1978) or extraction of the topicalized complement, leaving “an invisible resumptive pronoun” which is in turn extracted (Postal 1994); (ii) movement of a null operator (Chomsky 1977); and (iii) movements that can change category type (Stowell 1981; Pesetsky 1982; Webelhuth 1992). For criticisms and alternative analyses to the approaches in (iii) see Bresnan (1991, 1994b, 1995a), Hoekstra (1984), Plann (1986), Lujan´ (1994), and Neidle (1982b, 1988). For criticisms and alternatives to (ii) see Baltin (1982), Grimshaw (1993), Lasnik and Stowell (1991), Maling and Zaenen (1977), and Muller¨ and Sternefeld (1993). The problem with (i) is that in general, the kinds of phrases that can be topicalized do not coincide with the kinds of phrases that can bind resumptive pronouns. Quantifier phrases are one example discussed by Rizzi (1986), Rizzi and Roberts (1989, n. 1), Bresnan and Mchombo (1986, p. 279), Jelinek (1995), and Baker (1995), as illustrated by the Italian examples in (20):

(20) a. ∗Nessuno, lo conosco in questa citta.` Lit.: “Nobody, I know him in this city.” b. nessuno, conosco in questa citta.` “Nobody do I know in this city.”

9 Also see Dalrymple and Lødrup (2000) for discussion of variation in the grammatical function asso- ciated with that complements: obj versus comp. 20 Movement Paradoxes

There are other types of topicalized phrases which cannot bind resumptive definite pronouns (pace Postal 1994, pp. 67–68), as we see in English:

(21) a. An idiot he called me! Can you believe it? b. ∗An idiot he called me it! Can you believe it? (22) a. Last month I brought home a bold little kitten. rosie I named her. b. ∗rosie, I named her it.

It is true that a demonstrative pronoun can sometimes appear in these contexts, for example, Rosie – I named her that, but it requires a very different phrasing from examples (21) and (22): there must be a pause between the initial NP and the sentence, each having its own intonational contour. In (21) and (22), in con- trast, there is a single intonational contour for both the initial NP and the clause remnant, and the pitch on the NP is the prominence peak for the entire sentence. These facts suggest a different syntactic relation between the NP and the clause in the two cases. Lastly, English in general lacks resumptive pronouns in its gram- mar (Chao and Sells 1983; Sells 1984; Asudeh 2012), such that any resumption in that-clause topicalization would be quite special and would have to be very carefully motivated. See Asudeh (2012) for further discussion, including certain cases of topicalization which are argued not to involve grammatical resumption per se. Thus the resumptive pronoun solution to the category mismatch problem with topicalized that clauses lacks independent motivation and does not fit well with other facts of English grammar.