An Introduction to the Theory of Formal Languages and Automata

Total Page:16

File Type:pdf, Size:1020Kb

An Introduction to the Theory of Formal Languages and Automata JANUA LINGUARUM STUDIA MEMORIAE NICOLAI VAN WIJK DEDICATA edenda curat C. H. VAN SCHOONEVELD Indiana University Series Minor, 192/1 FORMAL GRAMMARS IN LINGUISTICS AND PSYCHOLINGUISTICS VOLUME I An Introduction to the Theory of Formal Languages and Automata by W. J. M. LEVELT 1974 MOUTON THE HAGUE • PARIS © Copyright 1974 in The Netherlands Mouton & Co. N. V., Publishers, The Hague No part of this book may be translated or reproduced in any form, by print, photoprint, microfilm, or any other means, without written permission from the publishers Translation: ANDREW BARNAS Printed in Belgium by N.I.C.I., Ghent PREFACE In the latter half of the 1950's, Noam Chomsky began to develop mathematical models for the description of natural languages. Two disciplines originated in his work and have grown to maturity. The first of these is the theory of formal grammars, a branch of mathematics which has proven to be of great interest to informa­ tion and computer sciences. The second is generative, or more specifically, transformational linguistics. Although these disciplines are independent and develop each according to its own aims and criteria, they remain closely interwoven. Without access to the theory of formal languages, for example, the contemporary study of the foundations of linguistics would be unthinkable. The collaboration of Chomsky and the psycholinguist, George Miller, around 1960 led to a considerable impact of transforma­ tional linguistics on the psychology of language. During a period of near feverish experimental activity, psycholinguists studied the various ways in which the new linguistic notions might be used in the development of models for language user and language acquisi­ tion. A good number of the original conceptions were naive and could not withstand critical test, but in spiteof this, transformational linguistics has greatly influenced modern psycholinguistics. The theory of formal languages, transformational linguistics, psycholinguistics, and their mutual relationships are the theme of this work. Volume I is an introduction to the theory of formal languages and automata; grammars are treated only as formal systems, and no application of the theory, linguistic or other, is made. Volume II in turn deals with applications of those mathe- VI PREFACE matical models to linguistic theory. Volume III treats applications of grammatical systems to models of language user and language learner, as well as the formal questions which have arisen as a result of such applications. The material is cumulative: Volume II supposes a general understanding of Volume I, and Volume III refers to the subjects dealt with in Volumes I and II. Volumes II and III have their own preface, so we can now turn to some introductory remarks with respect to the present volume. Volume I, independent of the two following volumes, should be seen as an introduction to the theory of formal languages and automata. A number of similar introductions are available at the moment, but I have nevertheless undertaken the present work for three reasons. First, most available texts, because they suppose an acquaintance with sophisticated mathematical theories and methods, are beyond the reach of many students of linguistics and psychology. More often than not, Chomsky's and Miller's contributions to the Handbook of Mathematical Psychology prove too difficult for early graduate teaching. The present introduction is kept at a rather elementary level; a general knowledge of college mathematics will be sufficient to follow the text, although familiarity with the elements of set theory and statistics will certainly be an advantage. Second, existing introductions treat a number of subjects which have little obvious relation to linguistics or psychology. The linguist or the psychologist is obliged to make his own selection from among a series of topics which he does not yet understand, and he might search in vain for a treatment of topics which are especially relevant to his field. Probabilistic grammars and gram­ matical inference, for example, are not treated in any of the existing introductions. Special attention has been paid to these topics in the present volume, but matters not directly relevant to linguistics or psychology have not been completely excluded, as a balanced presentation of the theory sets its own demands. The third reason for writing this introduction is to supply readers of the two following volumes with a concise survey of the main notions of formal language theory used there. The subject PREFACE VII index of this volume can be used to find definitions of technical terms: definitions are indicated by italicized page numbers. Without the help and cooperation of many, these three volumes could not have been realized. A first version was written during a sabbatical year at The Institute for Advanced Study in Princeton, New Jersey. I am deeply grateful to Professor Duncan Luce and to The Institute for the invitation which made my stay possible. Much in this work is due to the help and insights of Professor George Miller, former director of the Harvard Center for Cognitive Studies, where the new psychology of language originated under his guidance. Thanks to him I was granted a Research Fellowship at the Center in 1965, and by happy coincidence, he too was at the Institute for Advanced Study when I was composing the text. His attentive advice was most useful, especially in the writing of the third volume. Likewise, regular discussions with Dr. Philip Johnson- Laird helped to clarify many of the psychological issues. Conver­ sations with Professor Aravind Joshi on the subject matter of the first two volumes were also enormously stimulating and enjoyable; I profited almost daily from his erudition in the fields of both formal systems theory and mathematical linguistics. Finally, I wish to express my gratitude to all those who have contributed by critically reading the text in the original Dutch version: Professor L. Verbeek, Dr. H. Brandt Corstius, Mr. R. Brons, Dr. G. Kempen, Dr. A. van der Ven, Mr. E. Schils, Mr. L. Noordman, Dr. A. De Wachter-Schaerlaekens, and Professor A. Kraak. Their remarks not only prevented the printing of many disturbing errors, but also led to many enriching additions to the text. March 1973 W. J. M. Levelt Nijmegen TABLE OF CONTENTS Preface v 1. Grammars as Formal Systems 1 1.1. Grammars, Automata, and Inference 1 1.2. The Definition of "Grammar" 3 1.3. Examples 6 2. The Hierarchy of Grammars 9 2.1. Classes of Grammars 9 2.2. Regular Grammars 12 2.3. Context-free Grammars 16 2.3.1. The Chomsky Normal-Form 17 2.3.2. The Greibach Normal-Form 19 2.3.3. Self-embedding 21 2.3.4. Ambiguity 25 2.3.5. Linear Grammars 26 2.4. Context-sensitive Grammars 27 2.4.1. Context-sensitive productions 27 2.4.2. The Kuroda Normal-Form 31 3. Probabilistic Grammars 35 3.1. Definitions and Concepts 35 3.2. Classification 37 3.3. Regular Probabilistic Grammars 38 3.4. Context-free Probabilistic Grammars 44 3.4.1. Normal Forms 45 3.4.2. Consistency Conditions 50 X TABLE OF CONTENTS 4. Finite Automata 53 4.1. Definitions and Concepts 54 . 4.2. Nondeterministic Finite Automata 60 4.3. Finite Automata and Regular Grammars 63 4.4. Probabilistic Finite Automata 68 5. Push-Down Automata 75 5.1. Definitions and Concepts 76 5.2. Nondeterministic Push-down Automata and Context- free Languages 81 6. Linear-Bounded Automata 91 6.1. Definitions and Concepts 92 6.2. Linear-bounded Automata and Context-sensitive Languages 96 7. Turing Machines 101 7.1. Definitions and Concepts 102 7.2. A few Elementary Procedures 105 7.3. Turing Machines and Type-0 Languages 106 7.4. Mechanical Procedures, Recursive Enumerability, and Recursiveness 110 8. Grammatical Inference 115 8.1. Hypotheses, Observations, and Evaluation 115 8.2. The Classical Estimation of Parameters for Proba­ bilistic Grammars 118 8.3. The "Learnability" of Nonprobabilistic Languages . 121 8.4. Inference by means of Bayes' Theorem 124 Historical and Bibliographical Remarks 131 Bibliography 135 Author Index 139 Subject Index 140 1 GRAMMARS AS FORMAL SYSTEMS 1.1. GRAMMARS, AUTOMATA, AND INFERENCE The theory of formal languages originated in the study of natural languages. The description of a natural language is traditionally called a GRAMMAR; it should indicate how the sentences of a language are composed of elements, how elements form larger units, and how these units are related within the context of the sentence. The theory of formal languages proceeds from the need to provide a formal mathematical basis for such descriptions. Chomsky, the founder of the theory, envisaged more than a simple refinement of traditional linguistic description. He was primarily concerned with a more thorough examination of the basis of linguistic theory. This involves such questions as "what are the goals of linguistic theory?", "what conditions must a grammar fulfill in order to be adequate in view of these goals?", and "what is the general form of a linguistic theory?" Without a formal basis, these and similar questions cannot be handled with sufficient precision. Volume II of this book will deal with these issues; it will be shown that a formal language can serve as a mathematical model for a natural language, while a formal gram­ mar can act as a model for a linguistic theory. From a mathematical point of view, grammars are FORMAL SYSTEMS, like Turing machines, computer programs, prepositional logic, theories of inference, neural nets, and so forth. Formal systems characteristically transform a certain INPUT into a par­ ticular OUTPUT by means of completely explicit, mechanically applicable rules. Input and output are strings of symbols taken 2 GRAMMARS AS FORMAL SYSTEMS from a particular alphabet or VOCABULARY. For a formal grammar the input is an abstract START SYMBOL; the output is a string of "words" which constitutes a "sentence" of the formal "language".
Recommended publications
  • Formal Languages and Automata Theory Géza Horváth, Benedek Nagy Szerzői Jog © 2014 Géza Horváth, Benedek Nagy, University of Debrecen
    Formal Languages and Automata Theory Géza Horváth, Benedek Nagy Created by XMLmind XSL-FO Converter. Formal Languages and Automata Theory Géza Horváth, Benedek Nagy Szerzői jog © 2014 Géza Horváth, Benedek Nagy, University of Debrecen Created by XMLmind XSL-FO Converter. Tartalom Formal Languages and Automata Theory .......................................................................................... vi Introduction ...................................................................................................................................... vii 1. Elements of Formal Languages ...................................................................................................... 1 1. 1.1. Basic Terms .................................................................................................................... 1 2. 1.2. Formal Systems .............................................................................................................. 2 3. 1.3. Generative Grammars .................................................................................................... 5 4. 1.4. Chomsky Hierarchy ....................................................................................................... 6 2. Regular Languages and Finite Automata ...................................................................................... 12 1. 2.1. Regular Grammars ....................................................................................................... 12 2. 2.2. Regular Expressions ....................................................................................................
    [Show full text]
  • The Mathematics of Syntactic Structure: Trees and Their Logics
    Computational Linguistics Volume 26, Number 2 The Mathematics of Syntactic Structure: Trees and their Logics Hans-Peter Kolb and Uwe MÈonnich (editors) (Eberhard-Karls-UniversitÈat Tubingen) È Berlin: Mouton de Gruyter (Studies in Generative Grammar, edited by Jan Koster and Henk van Riemsdijk, volume 44), 1999, 347 pp; hardbound, ISBN 3-11-016273-3, $127.75, DM 198.00 Reviewed by Gerald Penn Bell Laboratories, Lucent Technologies Regular languages correspond exactly to those languages that can be recognized by a ®nite-state automaton. Add a stack to that automaton, and one obtains the context- free languages, and so on. Probably all of us learned at some point in our univer- sity studies about the Chomsky hierarchy of formal languages and the duality be- tween the form of their rewriting systems and the automata or computational re- sources necessary to recognize them. What is perhaps less well known is that yet another way of characterizing formal languages is provided by mathematical logic, namely in terms of the kind and number of variables, quanti®ers, and operators that a logical language requires in order to de®ne another formal language. This mode of characterization, which is subsumed by an area of research called descrip- tive complexity theory, is at once more declarative than an automaton or rewriting system, more ¯exible in terms of the primitive relations or concepts that it can pro- vide resort to, and less wedded to the tacit, not at all unproblematic, assumption that the right way to view any language, including natural language, is as a set of strings.
    [Show full text]
  • Grammars (Cfls & Cfgs) Reading: Chapter 5
    Context-Free Languages & Grammars (CFLs & CFGs) Reading: Chapter 5 1 Not all languages are regular So what happens to the languages which are not regular? Can we still come up with a language recognizer? i.e., something that will accept (or reject) strings that belong (or do not belong) to the language? 2 Context-Free Languages A language class larger than the class of regular languages Supports natural, recursive notation called “context- free grammar” Applications: Context- Parse trees, compilers Regular free XML (FA/RE) (PDA/CFG) 3 An Example A palindrome is a word that reads identical from both ends E.g., madam, redivider, malayalam, 010010010 Let L = { w | w is a binary palindrome} Is L regular? No. Proof: N N (assuming N to be the p/l constant) Let w=0 10 k By Pumping lemma, w can be rewritten as xyz, such that xy z is also L (for any k≥0) But |xy|≤N and y≠ + ==> y=0 k ==> xy z will NOT be in L for k=0 ==> Contradiction 4 But the language of palindromes… is a CFL, because it supports recursive substitution (in the form of a CFG) This is because we can construct a “grammar” like this: Same as: 1. A ==> A => 0A0 | 1A1 | 0 | 1 | Terminal 2. A ==> 0 3. A ==> 1 Variable or non-terminal Productions 4. A ==> 0A0 5. A ==> 1A1 How does this grammar work? 5 How does the CFG for palindromes work? An input string belongs to the language (i.e., accepted) iff it can be generated by the CFG G: Example: w=01110 A => 0A0 | 1A1 | 0 | 1 | G can generate w as follows: Generating a string from a grammar: 1.
    [Show full text]
  • Hierarchy and Interpretability in Neural Models of Language Processing ILLC Dissertation Series DS-2020-06
    Hierarchy and interpretability in neural models of language processing ILLC Dissertation Series DS-2020-06 For further information about ILLC-publications, please contact Institute for Logic, Language and Computation Universiteit van Amsterdam Science Park 107 1098 XG Amsterdam phone: +31-20-525 6051 e-mail: [email protected] homepage: http://www.illc.uva.nl/ The investigations were supported by the Netherlands Organization for Scientific Research (NWO), through a Gravitation Grant 024.001.006 to the Language in Interaction Consortium. Copyright © 2019 by Dieuwke Hupkes Publisher: Boekengilde Printed and bound by printenbind.nl ISBN: 90{6402{222-1 Hierarchy and interpretability in neural models of language processing Academisch Proefschrift ter verkrijging van de graad van doctor aan de Universiteit van Amsterdam op gezag van de Rector Magnificus prof. dr. ir. K.I.J. Maex ten overstaan van een door het College voor Promoties ingestelde commissie, in het openbaar te verdedigen op woensdag 17 juni 2020, te 13 uur door Dieuwke Hupkes geboren te Wageningen Promotiecommisie Promotores: Dr. W.H. Zuidema Universiteit van Amsterdam Prof. Dr. L.W.M. Bod Universiteit van Amsterdam Overige leden: Dr. A. Bisazza Rijksuniversiteit Groningen Dr. R. Fern´andezRovira Universiteit van Amsterdam Prof. Dr. M. van Lambalgen Universiteit van Amsterdam Prof. Dr. P. Monaghan Lancaster University Prof. Dr. K. Sima'an Universiteit van Amsterdam Faculteit der Natuurwetenschappen, Wiskunde en Informatica to my parents Aukje and Michiel v Contents Acknowledgments xiii 1 Introduction 1 1.1 My original plan . .1 1.2 Neural networks as explanatory models . .2 1.2.1 Architectural similarity . .3 1.2.2 Behavioural similarity .
    [Show full text]
  • The Complexity of Narrow Syntax : Minimalism, Representational
    Erschienen in: Measuring Grammatical Complexity / Frederick J. Newmeyer ... (Hrsg.). - Oxford : Oxford University Press, 2014. - S. 128-147. - ISBN 978-0-19-968530-1 The complexity of narrow syntax: Minimalism, representational economy, and simplest Merge ANDREAS TROTZKE AND JAN-WOUTER ZWART 7.1 Introduction The issue of linguistic complexity has recently received much attention by linguists working within typological-functional frameworks (e.g. Miestamo, Sinnemäki, and Karlsson 2008; Sampson, Gil, and Trudgill 2009). In formal linguistics, the most prominent measure of linguistic complexity is the Chomsky hierarchy of formal languages (Chomsky 1956), including the distinction between a finite-state grammar (FSG) and more complicated types of phrase-structure grammar (PSG). This dis- tinction has played a crucial role in the recent biolinguistic literature on recursive complexity (Sauerland and Trotzke 2011). In this chapter, we consider the question of formal complexity measurement within linguistic minimalism (cf. also Biberauer et al., this volume, chapter 6; Progovac, this volume, chapter 5) and argue that our minimalist approach to complexity of derivations and representations shows simi- larities with that of alternative theoretical perspectives represented in this volume (Culicover, this volume, chapter 8; Jackendoff and Wittenberg, this volume, chapter 4). In particular, we agree that information structure properties should not be encoded in narrow syntax as features triggering movement, suggesting that the relevant information is established at the interfaces. Also, we argue for a minimalist model of grammar in which complexity arises out of the cyclic interaction of subderivations, a model we take to be compatible with Construction Grammar approaches. We claim that this model allows one to revisit the question of the formal complexity of a generative grammar, rephrasing it such that a different answer to the question of formal complexity is forthcoming depending on whether we consider the grammar as a whole, or just narrow syntax.
    [Show full text]
  • Formal Language and Automata Theory (CS21004)
    Context Free Grammar Normal Forms Derivations and Ambiguities Pumping lemma for CFLs PDA Parsing CFL Properties Formal Language and Automata Theory (CS21004) Soumyajit Dey CSE, IIT Kharagpur Context Free Formal Language and Automata Theory Grammar (CS21004) Normal Forms Derivations and Ambiguities Pumping lemma Soumyajit Dey for CFLs CSE, IIT Kharagpur PDA Parsing CFL Properties DPDA, DCFL Membership CSL Soumyajit Dey CSE, IIT Kharagpur Formal Language and Automata Theory (CS21004) Context Free Grammar Normal Forms Derivations and Ambiguities Pumping lemma for CFLs PDA Parsing CFL Properties Formal Language and Automata Announcements Theory (CS21004) Soumyajit Dey CSE, IIT Kharagpur Context Free Grammar Normal Forms Derivations and The slide is just a short summary Ambiguities Follow the discussion and the boardwork Pumping lemma for CFLs Solve problems (apart from those we dish out in class) PDA Parsing CFL Properties DPDA, DCFL Membership CSL Soumyajit Dey CSE, IIT Kharagpur Formal Language and Automata Theory (CS21004) Context Free Grammar Normal Forms Derivations and Ambiguities Pumping lemma for CFLs PDA Parsing CFL Properties Formal Language and Automata Table of Contents Theory (CS21004) Soumyajit Dey 1 Context Free Grammar CSE, IIT Kharagpur 2 Normal Forms Context Free Grammar 3 Derivations and Ambiguities Normal Forms 4 Derivations and Pumping lemma for CFLs Ambiguities 5 Pumping lemma PDA for CFLs 6 Parsing PDA Parsing 7 CFL Properties CFL Properties DPDA, DCFL 8 DPDA, DCFL Membership 9 Membership CSL 10 CSL Soumyajit Dey CSE,
    [Show full text]
  • What Was Wrong with the Chomsky Hierarchy?
    What Was Wrong with the Chomsky Hierarchy? Makoto Kanazawa Hosei University Chomsky Hierarchy of Formal Languages recursively enumerable context- sensitive context- free regular Enduring impact of Chomsky’s work on theoretical computer science. 15 pages devoted to the Chomsky hierarchy. Hopcroft and Ullman 1979 No mention of the Chomsky hierarchy. 450 INDEX The same with the latest edition of Carmichael, R. D., 444 CNF-formula, 302 Cartesian product, 6, 46 Co-Turing-recognizableHopcroft, language, 209 Motwani, and Ullman (2006). CD-ROM, 349 Cobham, Alan, 444 Certificate, 293 Coefficient, 183 CFG, see Context-free grammar Coin-flip step, 396 CFL, see Context-free language Complement operation, 4 Chaitin, Gregory J., 264 Completed rule, 140 Chandra, Ashok, 444 Complexity class Characteristic sequence, 206 ASPACE(f(n)),410 Checkers, game of, 348 ATIME(t(n)),410 Chernoff bound, 398 BPP,397 Chess, game of, 348 coNL,354 Chinese remainder theorem, 401 coNP,297 Chomsky normal form, 108–111, 158, EXPSPACE,368 198, 291 EXPTIME,336 Chomsky, Noam, 444 IP,417 Church, Alonzo, 3, 183, 255 L,349 Church–Turing thesis, 183–184, 281 NC,430 CIRCUIT-SAT,386 NL,349 Circuit-satisfiability problem, 386 NP,292–298 CIRCUIT-VALUE,432 NPSPACE,336 Circular definition, 65 NSPACE(f(n)),332 Clause, 302 NTIME(f(n)),295 Clique, 28, 296 P,284–291,297–298 CLIQUE,296 PH,414 Sipser 2013Closed under, 45 PSPACE,336 Closure under complementation RP,403 context-free languages, non-, 154 SPACE(f(n)),332 deterministic context-free TIME(f(n)),279 languages, 133 ZPP,439 P,322 Complexity
    [Show full text]
  • INF2080 Context-Sensitive Langugaes
    INF2080 Context-Sensitive Langugaes Daniel Lupp Universitetet i Oslo 15th February 2017 Department of University of Informatics Oslo INF2080 Lecture :: 15th February 1 / 20 Context-Free Grammar Definition (Context-Free Grammar) A context-free grammar is a 4-tuple (V ; Σ; R; S) where 1 V is a finite set of variables 2 Σ is a finite set disjoint from V of terminals 3 R is a finite set of rules, each consisting of a variable and of a string of variables and terminals 4 and S is the start variable Rules are of the form A ! B1B2B3 ::: Bm, where A 2 V and each Bi 2 V [ Σ. INF2080 Lecture :: 15th February 2 / 20 consider the following toy programming language, where you can “declare” and “assign” a variable a value. S ! declare v; S j assign v : x; S this is context-free... but what if we only want to allow assignment after declaration and an infinite amount of variable names? ! context-sensitive! Why Context-Sensitive? Many building blocks of programming languages are context-free, but not all! INF2080 Lecture :: 15th February 3 / 20 this is context-free... but what if we only want to allow assignment after declaration and an infinite amount of variable names? ! context-sensitive! Why Context-Sensitive? Many building blocks of programming languages are context-free, but not all! consider the following toy programming language, where you can “declare” and “assign” a variable a value. S ! declare v; S j assign v : x; S INF2080 Lecture :: 15th February 3 / 20 but what if we only want to allow assignment after declaration and an infinite amount of variable names? ! context-sensitive! Why Context-Sensitive? Many building blocks of programming languages are context-free, but not all! consider the following toy programming language, where you can “declare” and “assign” a variable a value.
    [Show full text]
  • Automata and Grammars
    Automata and Grammars Prof. Dr. Friedrich Otto Fachbereich Elektrotechnik/Informatik, Universität Kassel 34109 Kassel, Germany E-mail: [email protected] SS 2018 Prof. Dr. F. Otto (Universität Kassel) Automata and Grammars 1 / 294 Lectures and Seminary SS 2018 Lectures: Thursday 9:00 - 10:30, Room S 11 Start: Thursday, February 22, 2018, 9:00. Seminary: Thursday 10:40 - 12:10, Room S 11 Start: Thursday, February 22, 2018, 10:40. Prof. Dr. F. Otto (Universität Kassel) Automata and Grammars 2 / 294 Exercises: From Thursday to Thursday next week! To pass seminary: At least two successful presentations in class (10 %) and passing of midtem exam on April 5 (10:40 - 11:40) (20 %) Final Exam: (dates have been fixed!) Written exam (2 hours) on May 31 (9:00 - 11:30) in S 11 Oral exam (30 minutes per student) on June 14 (9:00 - 12:30) in S 10 Written exam (2. try) on June 21 (9:00 - 11:30) in S 11 Oral exam (2. try) on June 28 (9:00 - 12:30) in S 11 The written exam accounts for 50 % of the final grade, while the oral exam accounts for 20 % of the final grade. Moodle: Course ”Automata and Grammars” NTIN071. Prof. Dr. F. Otto (Universität Kassel) Automata and Grammars 3 / 294 Literature: P.J. Denning, J.E. Dennis, J.E. Qualitz; Machines, Languages, and Computation. Prentice-Hall, Englewood Cliffs, N.J., 1978. M. Harrison; Introduction to Formal Language Theory. Addison-Wesley, Reading, M.A., 1978. J.E. Hopcroft, R. Motwani, J.D. Ullman; Introduction to Automata Theory, Languages, and Computation.
    [Show full text]
  • Reflection in the Chomsky Hierarchy
    Journal of Automata, Languages and Combinatorics 18 (2013) 1, 53–60 c Otto-von-Guericke-Universit¨at Magdeburg REFLECTION IN THE CHOMSKY HIERARCHY Henk Barendregt Institute of Computing and Information Science, Radboud University, Nijmegen, The Netherlands e-mail: [email protected] Venanzio Capretta School of Computer Science, University of Nottingham, United Kingdom e-mail: [email protected] and Dexter Kozen Computer Science Department, Cornell University, Ithaca, New York 14853, USA e-mail: [email protected] ABSTRACT The class of regular languages can be generated from the regular expressions. These regular expressions, however, do not themselves form a regular language, as can be seen using the pumping lemma. On the other hand, the class of enumerable languages can be enumerated by a universal language that is one of its elements. We say that the enumerable languages are reflexive. In this paper we investigate what other classes of the Chomsky Hierarchy are re- flexive in this sense. To make this precise we require that the decoding function is itself specified by a member of the same class. Could it be that the regular languages are reflexive, by using a different collection of codes? It turns out that this is impossi- ble: the collection of regular languages is not reflexive. Similarly the collections of the context-free, context-sensitive, and computable languages are not reflexive. Therefore the class of enumerable languages is the only reflexive one in the Chomsky Hierarchy. In honor of Roel de Vrijer on the occasion of his sixtieth birthday 1. Introduction Let (Σ∗) be a class of languages over an alphabet Σ.
    [Show full text]
  • Noam Chomsky
    Noam Chomsky “If you’re teaching today what you were teaching five years ago, either the field is dead or you are.” Background ▪ Jewish American born to Ukranian and Belarussian immigrants. ▪ Attended a non-competitive elementary school. ▪ Considered dropping out of UPenn and moving to British Palestine. ▪ Life changed when introduced to linguist Zellig Harris. Contributions to Cognitive Science • Deemed the “father of modern linguistics” • One of the founders of Cognitive Science o Linguistics: Syntactic Structures (1957) o Universal Grammar Theory o Generative Grammar Theory o Minimalist Program o Computer Science: Chomsky Hierarchy o Philosophy: Philosophy of Mind; Philosophy of Language o Psychology: Critical of Behaviorism Transformational Grammar – 1955 Syntactic Structures - 1957 ▪ Made him well-known within the ▪ Big Idea: Given the grammar, linguistics community you should be able to construct various expressions in the ▪ Revolutionized the study of natural language with prior language knowledge. ▪ Describes generative grammars – an explicit statement of what the classes of linguistic expressions in a language are and what kind of structures they have. Universal Grammar Theory ▪ Credited with uncovering ▪ Every sentence has a deep structure universal properties of natural that is mapped onto the surface human languages. structure according to rules. ▪ Key ideas: Humans have an ▪ Language similarities: innate neural capacity for – Deep Structure: Pattern in the mind generating a system of rules. – Surface Structure: Spoken utterances – There is a universal grammar that – Rules: Transformations underlies ALL languages Chomsky Hierarchy ▪ Hierarchy of classes of formal grammars ▪ Focuses on structure of classes and relationship between grammars ▪ Typically used in Computer Science and Linguistics Minimalist Program – 1993 ▪ Program NOT theory ▪ Provides conceptual framework to guide development of linguistic theory.
    [Show full text]
  • The Evolution of the Faculty of Language from a Chomskyan Perspective: Bridging Linguistics and Biology
    doi 10.4436/JASS.91011 JASs Invited Reviews e-pub ahead of print Journal of Anthropological Sciences Vol. 91 (2013), pp. 1-48 The evolution of the Faculty of Language from a Chomskyan perspective: bridging linguistics and biology Víctor M. Longa Area of General Linguistics, Faculty of Philology, University of Santiago de Compostela, 15782 Santiago de Compostela, Spain e-mail: [email protected] Summary - While language was traditionally considered a purely cultural trait, the advent of Noam Chomsky’s Generative Grammar in the second half of the twentieth century dramatically challenged that view. According to that theory, language is an innate feature, part of the human biological endowment. If language is indeed innate, it had to biologically evolve. This review has two main objectives: firstly, it characterizes from a Chomskyan perspective the evolutionary processes by which language could have come into being. Secondly, it proposes a new method for interpreting the archaeological record that radically differs from the usual types of evidence Paleoanthropology has concentrated on when dealing with language evolution: while archaeological remains have usually been regarded from the view of the behavior they could be associated with, the paper will consider archaeological remains from the view of the computational processes and capabilities at work for their production. This computational approach, illustrated with a computational analysis of prehistoric geometric engravings, will be used to challenge the usual generative thinking on language evolution, based on the high specificity of language. The paper argues that the biological machinery of language is neither specifically linguistic nor specifically human, although language itself can still be considered a species-specific innate trait.
    [Show full text]