Indexed Languages (And Why You Care)

Total Page:16

File Type:pdf, Size:1020Kb

Indexed Languages (And Why You Care) Indexed Languages and why you care Presented by Pieter Hooimeijer 2008-02-07 1 Think Way Back... 2 www.xkcd.com ...Far Enough: ... CS NCF DCF REG REG 3 Formal Language Classes ... CS NCF DCF REG REG PDA 4 Formal Language Classes ... CS NCF DCF REG REG PDA DFA 5 Formal Language Classes ... CS NCF DCF REG REG Machines PDA DFA Formalism S → aSb | z a(ab)*b 6 Formal Language Classes ... CS NCF DCF REG REG Pump Pump 7 The Context-Free Pumping Lemma Suppose L = { anbncn | n = 1...} is context- free. By the pumping lemma, any string s with |s| ≥ p can be 'pumped.' p 8 The Context-Free Pumping Lemma What does 'can be pumped' mean? s = uvxyz 1)|vy| > 0 2)|vxy| ≤ p i i 3) uv xy z is in L for all i ≥ 0 9 The Context-Free Pumping Lemma Suppose L = { anbncn | n = 1...} is context-free. Consider s = apbpcp; for any split s = uvxyz, we have: - if v contains a's, then y cannot contain c's - if y contains c's, then v cannot contain a's 0 0 - a problem: uv xy z = uxz will never be in L 10 Moving Right Along 11 Motivation – Some Examples • Can solve problems by phrasing them as 'language' problems: • Finding valid control flow graph paths • Solving set constraint problems • Static string analysis • Lexing, parsing • For fun... 12 Example – String Variables www.xkcd.com 13 The Nugget Some Code: x = 'z'; • We want a context free grammar to model x while(n < 5) { x = '(' . x . ')'; n ++; • Suppose we don't know } anything about n 14 The Nugget Some Code: Grammar: >x = 'z'; A -> z while(n < 5) { x = '(' . x . ')'; n ++; } 15 The Nugget Some Code: Grammar: x = 'z'; A -> z [True] while(n < 5) { B -> (A) [n < 5] > x = '(' . x . ')'; n ++; } 16 The Nugget Some Code: Grammar: x = 'z'; A -> z [True] while(n < 5) { B -> (A) [n < 5] x = '(' . x . ')'; C -> A | B [n < 5] n ++; >>} 17 The Nugget Some Code: Grammar: x = 'z'; A -> z [True] B -> (C) [n < 5] while(n < 5) { C -> A | B [True] x = '(' . x . ')'; n ++; >>} 18 The Nugget Some Code: Grammar: x = 'z'; X -> C while(n < 5) { A -> z [True] x = '(' . x . ')'; B -> (C) [n < 5] n ++; C -> A | B [True] >>} 19 Indexed Languages ... CS *I* NCF DCF REG REG 20 Definition: Indexed Grammar • G = (N, T, F, P, S) P : { N → ((NF*) ∪ T)* } F : { { N → (N ∪ T)*}f } 21 Derivation Rule Example S 1 S → Sf S → ABgC Sf Sff Aff Bgff Cff 22 Derivation Rule Example S 1 S → Sf S → ABgC Sf Sff 'normal' production: copy index stack Aff Bgff Cff 23 Derivation Rule Example S 2 S → Af F = { { A → B}f } Af B● 24 Derivation Rule Example S 2 S → Af F = { { A → B}f } Af B 25 Derivation Rule Example S 2 S → Af F = { { A → B}f } Af B 'index' production: pop leftmost index (for current nonterminal only) 26 An Example Grammar S → Df g = {A → Aa f = {A → a D → Dg | ABC B → Bb B → b C → Cc} C → c} 27 An Example Grammar S → Df g = {A → Aa f = {A → a D → Dg | ABC B → Bb B → b C → Cc} C → c} What is the language of this grammar? 28 An Example Grammar S → Df S D → Dg | ABC g = {A → Aa B → Bb C → Cc} f = {A → a B → b C → c} 29 An Example Grammar S → Df S D → Dg | ABC Df g = {A → Aa B → Bb C → Cc} f = {A → a B → b C → c} 30 An Example Grammar S → Df S D → Dg | ABC Df g = {A → Aa Dgf B → Bb C → Cc} f = {A → a B → b C → c} 31 An Example Grammar S → Df S D → Dg | ABC Df g = {A → Aa Dgf B → Bb C → Cc} Agf Bgf Cgf f = {A → a B → b C → c} 32 An Example Grammar S → Df S D → Dg | ABC Df g = {A → Aa Dgf B → Bb C → Cc} Agf Bgf Cgf f = {A → a B → b C → c} 33 An Example Grammar S → Df S D → Dg | ABC Df g = {A → Aa Dgf B → Bb C → Cc} Agf Bgf Cgf f = {A → a Afa B → b C → c} 34 An Example Grammar S → Df S D → Dg | ABC Df g = {A → Aa Dgf B → Bb C → Cc} Agf Bgf Cgf f = {A → a Afa a B → b C → c} aa 35 An Example Grammar S → Df S D → Dg | ABC Df g = {A → Aa Dgf B → Bb C → Cc} Agf Bgf Cgf Afa f = {A → a Bfb a B → b b C → c} aa bb 36 An Example Grammar S → Df S D → Dg | ABC Df g = {A → Aa Dgf B → Bb C → Cc} Agf Bgf Cgf Afa Cfc f = {A → a Bfb a c B → b b C → c} aa bb cc 37 Where am I? { anbncn | n = 1...} ... CS *I* NCF DCF REG REG 38 ... CS *I* NCF DCF REG REG Machines PDA DFA Formalisms S → aSb | z a(ab)*b Pumps 39 ... CS *I* NCF DCF REG REG NSA PDA DFA S → aSfb | z S → aSb | z a(ab)*b ? 40 Associated Automaton 41 Associated Automaton 42 (Actual NSA) [Gilman, Shapiro 1998] 43 But will it pump? • It's complicated – on derivation trees rather than strings [Hayashi 1973] • Several corollaries exist [Gilman 1996] (less general; easier to use) 44 The Lemma L : indexed language m : positive integer There is a constant k > 0 so that each w in L can be split (w = w1w2...wr) subject to: The Lemma L : indexed language m : positive integer There is a constant k > 0 so that each w in L can be split (w = w1w2...wr) subject to: • m < r ≤ k • each |wi| > 0 The Lemma L : indexed language m : positive integer There is a constant k > 0 so that each w in L can be split (w = w1w2...wr) subject to: • m < r ≤ k • each |wi| > 0 • any m-sized set of wi's is a subset of some w' in L; w' is a subproduct of w Lemma Example Lemma Example (Montage Time) 50 { (abn)n | n = 1...} { anbncn | n = 1...} ... CS *I* NCF DCF REG REG NSA PDA DFA S → aSfb | z S → aSb | z a(ab)*b ? 51 52 References • Indexed Grammars – An Extension to Context-Free Grammars (Aho) • Nested Stack Automata (Aho) • Sequentially Indexed Grammars (van Eijck) • A Shrinking Lemma for Indexed Languages (Gilman) • On Groups Whose Word Problem is Solved by a Nested Stack Automaton (Gilman and Shapiro) • On Derivation Trees of Indexed Grammars (Hayashi) 53.
Recommended publications
  • The Structure of Index Sets and Reduced Indexed Grammars Informatique Théorique Et Applications, Tome 24, No 1 (1990), P
    INFORMATIQUE THÉORIQUE ET APPLICATIONS R. PARCHMANN J. DUSKE The structure of index sets and reduced indexed grammars Informatique théorique et applications, tome 24, no 1 (1990), p. 89-104 <http://www.numdam.org/item?id=ITA_1990__24_1_89_0> © AFCET, 1990, tous droits réservés. L’accès aux archives de la revue « Informatique théorique et applications » im- plique l’accord avec les conditions générales d’utilisation (http://www.numdam. org/conditions). Toute utilisation commerciale ou impression systématique est constitutive d’une infraction pénale. Toute copie ou impression de ce fichier doit contenir la présente mention de copyright. Article numérisé dans le cadre du programme Numérisation de documents anciens mathématiques http://www.numdam.org/ Informatique théorique et Applications/Theoretical Informaties and Applications (vol. 24, n° 1, 1990, p. 89 à 104) THE STRUCTURE OF INDEX SETS AND REDUCED INDEXED GRAMMARS (*) by R. PARCHMANN (*) and J. DUSKE (*) Communicated by J. BERSTEL Abstract. - The set of index words attached to a variable in dérivations of indexed grammars is investigated. Using the regularity of these sets it is possible to transform an mdexed grammar in a reducedfrom and to describe the structure ofleft sentential forms of an indexed grammar. Résumé. - On étudie Vensemble des mots d'index d'une variable dans les dérivations d'une grammaire d'index. La rationalité de ces ensembles peut être utilisée pour transformer une gram- maire d'index en forme réduite, et pour décrire la structure des mots apparaissant dans les dérivations gauches d'une grammaire d'index. 1. INTRODUCTION In this paper we will further investigate indexed grammars and languages introduced by Aho [1] as an extension of context-free grammars and lan- guages.
    [Show full text]
  • Genome Grammars
    Genome Grammars Genome Sequences and Formal Languages Andreas de Vries Version: June 17, 2011 Wir sind aus Staub und Fantasie Andreas Bourani, Nur in meinem Kopf (2011) Contents 1 Genetics 5 1.1 Cell physiology ............................. 5 1.2 Amino acids and proteins ....................... 6 1.2.1 Geometry of peptide bonds .................. 8 1.2.2 Protein structure ......................... 10 1.3 Nucleic acids ............................... 12 1.4 DNA replication ............................. 14 1.5 Flow of information for cell growth .................. 16 1.5.1 The genetic code ......................... 18 1.5.2 Open reading frames, coding regions, and genes ...... 19 2 Formal languages 23 2.1 The Chomsky hierarchy ......................... 25 2.2 Regular languages ............................ 28 2.2.1 Regular expressions ....................... 30 2.3 Context-free languages ......................... 31 2.3.1 Linear languages ........................ 33 2.4 Context-sensitive languages ...................... 34 2.4.1 Indexed languages ....................... 35 2.5 Languages and machines ........................ 38 3 Grammar of DNA Strings 42 3.1 Searl’s approach to DNA language .................. 42 3.2 Gene regulation and inadequacy of context-free grammars ..... 45 3.3 DNA splicing rule ............................ 46 A Mathematical Foundations 49 A.1 Notation ................................. 49 A.2 Sets .................................... 49 A.3 Maps ................................... 52 A.4 Algebra .................................
    [Show full text]
  • What Was Wrong with the Chomsky Hierarchy?
    What Was Wrong with the Chomsky Hierarchy? Makoto Kanazawa Hosei University Chomsky Hierarchy of Formal Languages recursively enumerable context- sensitive context- free regular Enduring impact of Chomsky’s work on theoretical computer science. 15 pages devoted to the Chomsky hierarchy. Hopcroft and Ullman 1979 No mention of the Chomsky hierarchy. 450 INDEX The same with the latest edition of Carmichael, R. D., 444 CNF-formula, 302 Cartesian product, 6, 46 Co-Turing-recognizableHopcroft, language, 209 Motwani, and Ullman (2006). CD-ROM, 349 Cobham, Alan, 444 Certificate, 293 Coefficient, 183 CFG, see Context-free grammar Coin-flip step, 396 CFL, see Context-free language Complement operation, 4 Chaitin, Gregory J., 264 Completed rule, 140 Chandra, Ashok, 444 Complexity class Characteristic sequence, 206 ASPACE(f(n)),410 Checkers, game of, 348 ATIME(t(n)),410 Chernoff bound, 398 BPP,397 Chess, game of, 348 coNL,354 Chinese remainder theorem, 401 coNP,297 Chomsky normal form, 108–111, 158, EXPSPACE,368 198, 291 EXPTIME,336 Chomsky, Noam, 444 IP,417 Church, Alonzo, 3, 183, 255 L,349 Church–Turing thesis, 183–184, 281 NC,430 CIRCUIT-SAT,386 NL,349 Circuit-satisfiability problem, 386 NP,292–298 CIRCUIT-VALUE,432 NPSPACE,336 Circular definition, 65 NSPACE(f(n)),332 Clause, 302 NTIME(f(n)),295 Clique, 28, 296 P,284–291,297–298 CLIQUE,296 PH,414 Sipser 2013Closed under, 45 PSPACE,336 Closure under complementation RP,403 context-free languages, non-, 154 SPACE(f(n)),332 deterministic context-free TIME(f(n)),279 languages, 133 ZPP,439 P,322 Complexity
    [Show full text]
  • Iterated Stack Automata and Complexity Classes
    INFORMATION AND COMPUTATION 95, 2 1-75 ( t 99 1) Iterated Stack Automata and Complexity Classes JOOST ENCELFRIET Department of Computer Science, Leiden University. P.O. Box 9.512, 2300 RA Leiden, The Netherlands An iterated pushdown is a pushdown of pushdowns of . of pushdowns. An iterated exponential function is 2 to the 2 to the to the 2 to some polynomial. The main result presented here is that the nondeterministic 2-way and multi-head iterated pushdown automata characterize the deterministic iterated exponential time complexity classes. This is proved by investigating both nondeterministic and alternating auxiliary iterated pushdown automata, for which similar characteriza- tion results are given. In particular it is shown that alternation corresponds to one more iteration of pushdowns. These results are applied to the l-way iterated pushdown automata: (1) they form a proper hierarchy with respect to the number of iterations, and (2) their emptiness problem is complete in deterministic iterated exponential time. Similar results are given for iterated stack (checking stack, non- erasing stack, nested stack, checking stack-pushdown) automata. ? 1991 Academic Press. Inc INTRODUCTION It is well known that several types of 2-way and multi-head pushdown automata and stack automata have the same power as certain time or space bounded Turing machines; see, e.g., Chapter 14 of (Hopcroft and Ullman, 1979), or Sections 13 and 20.2 of (Wagner and Wechsung, 1986). For the deterministic and nondeterministic case such characterizations were given by Fischer (1969) for checking stack automata, by Hopcroft and Ullman (1967b) for nonerasing stack automata, by Cook (1971) for auxiliary pushdown and 2-way stack automata, by Ibarra (1971) for auxiliary (nonerasing and erasing) stack automata, by Beeri (1975) for 2-way and auxiliary nested stack automata, and by van Leeuwen (1976) for auxiliary checking stack-pushdown automata.
    [Show full text]
  • Hairdressing in Groups: a Survey of Combings and Formal Languages 1
    ISSN 1464-8997 493 Geometry & Topology Monographs Volume 1: The Epstein Birthday Schrift Pages 493–509 Hairdressing in groups: a survey of combings and formal languages Sarah Rees Abstract A group is combable if it can be represented by a language of words satisfying a fellow traveller property; an automatic group has a synchronous combing which is a regular language. This article surveys results for combable groups, in particular in the case where the combing is a formal language. AMS Classification 20F10, 20-04, 68Q40; 03D40 Keywords Combings, formal languages, fellow travellers, automatic groups Dedicated to David Epstein on the occasion of his 60th birthday 1 Introduction The aim of this article is to survey work generalising the notion of an automatic group, in particular to classes of groups associated with various classes of formal languages in the same way that automatic groups are associated with regular languages. The family of automatic groups, originally defined by Thurston in an attempt to abstract certain finiteness properties of the fundamental groups of hyperbolic manifolds recognised by Cannon in [12], has been of interest for some time. The defining properties of the family give a geometrical viewpoint on the groups and facilitate computation with them; to such a group is associated a set of paths in the Cayley graph of the group (a ‘language’ for the group) which both satisfies a geometrical ‘fellow traveller condition’ and, when viewed as a set of words, lies in the formal language class of regular languages. (A formal definition is given in section 2.) Epstein et al.’s book [15] gives a full account; the papers [3] and [16] are also useful references (in particular, [16] is very readable and non-technical).
    [Show full text]
  • Mildly Context-Sensitive Grammar-Formalisms
    Mildly Context Sensitive Grammar Formalisms Petra Schmidt Pfarrer-Lauer-Strasse 23 66386 St. Ingbert [email protected] Overview ,QWURGXFWLRQ«««««««««««««««««««««««««««««««««««« 3 Tree-Adjoining-Grammars 7$*V «««««««««««««««««««««««««« 4 CoPELQDWRU\&DWHJRULDO*UDPPDU &&*V ««««««««««««««««««««««« 7 /LQHDU,QGH[HG*UDPPDU /,*V ««««««««««««««««««««««««««« 8 3URRIRIHTXLYDOHQFH«««««««««««««««««««««««««««««««« 9 CCG ĺ /,*«««««««««««««««««««««««««««««««««« 9 LIG ĺ 7$*«««««««««««««««««««««««««««««««««« 10 TAG ĺ &&*«««««««««««««««««««««««««««««««««« 13 Linear Context-)UHH5HZULWLQJ6\VWHPV /&)56V «««««««««««««««««««« 15 Multiple Context-Free-*UDPPDUV««««««««««««««««««««««««««« 16 &RQFOXVLRQ«««««««««««««««««««««««««««««««««««« 17 References««««««««««««««««««««««««««««««««««««« 18 Introduction Natural languages are not context free. An example non-context-free phenomenon are cross-serial dependencies in Dutch, an example of which we show in Fig. 1. ik haar hem de nijlpaarden zag helpen voeren Fig. 1.: cross-serial dependencies A concept motivated by the intention of characterizing a class of formal grammars, which allow the description of natural languages such as Dutch with its cross-serial dependencies was introduced in 1985 by Aravind Joshi. This class of formal grammars is known as the class of mildly context-sensitive grammar-formalisms. According to Joshi (1985, p.225), a mildly context-sensitive language L has to fulfill three criteria to be understood as a rough characterization. These are: The parsing problem for L is solvable in
    [Show full text]
  • Cs.FL] 7 Dec 2020 4 a Model Programming Language and Its Grammar 11 4.1 Alphabet
    Describing the syntax of programming languages using conjunctive and Boolean grammars∗ Alexander Okhotin† December 8, 2020 Abstract A classical result by Floyd (\On the non-existence of a phrase structure grammar for ALGOL 60", 1962) states that the complete syntax of any sensible programming language cannot be described by the ordinary kind of formal grammars (Chomsky's \context-free"). This paper uses grammars extended with conjunction and negation operators, known as con- junctive grammars and Boolean grammars, to describe the set of well-formed programs in a simple typeless procedural programming language. A complete Boolean grammar, which defines such concepts as declaration of variables and functions before their use, is constructed and explained. Using the Generalized LR parsing algorithm for Boolean grammars, a pro- gram can then be parsed in time O(n4) in its length, while another known algorithm allows subcubic-time parsing. Next, it is shown how to transform this grammar to an unambiguous conjunctive grammar, with square-time parsing. This becomes apparently the first specifica- tion of the syntax of a programming language entirely by a computationally feasible formal grammar. Contents 1 Introduction 2 2 Conjunctive and Boolean grammars 5 2.1 Conjunctive grammars . .5 2.2 Boolean grammars . .6 2.3 Ambiguity . .7 3 Language specification with conjunctive and Boolean grammars 8 arXiv:2012.03538v1 [cs.FL] 7 Dec 2020 4 A model programming language and its grammar 11 4.1 Alphabet . 11 4.2 Lexical conventions . 12 4.3 Identifier matching . 13 4.4 Expressions . 14 4.5 Statements . 15 4.6 Function declarations . 16 4.7 Declaration of variables before use .
    [Show full text]
  • From Indexed Grammars to Generating Functions∗
    RAIRO-Theor. Inf. Appl. 47 (2013) 325–350 Available online at: DOI: 10.1051/ita/2013041 www.rairo-ita.org FROM INDEXED GRAMMARS TO GENERATING FUNCTIONS ∗ Jared Adams1, Eric Freden1 and Marni Mishna2 Abstract. We extend the DSV method of computing the growth series of an unambiguous context-free language to the larger class of indexed languages. We illustrate the technique with numerous examples. Mathematics Subject Classification. 68Q70, 68R15. 1. Introduction 1.1. Indexed grammars Indexed grammars were introduced in the thesis of Aho in the late 1960s to model a natural subclass of context-sensitive languages, more expressive than context-free grammars with interesting closure properties [1,16]. The original ref- erence for basic results on indexed grammars is [1]. The complete definition of these grammars is equivalent to the following reduced form. Definition 1.1. A reduced indexed grammar is a 5-tuple (N , T , I, P, S), such that 1. N , T and I are three mutually disjoint finite sets of symbols: the set N of non-terminals (also called variables), T is the set of terminals and I is the set of indices (also called flags); 2. S ∈N is the start symbol; 3. P is a finite set of productions, each having the form of one of the following: (a) A → α Keywords and phrases. Indexed grammars, generating functions, functional equations, DSV method. ∗ The third author gratefully acknowledges the support of NSERC Discovery Grant funding (Canada), and LaBRI (Bordeaux) for hosting during the completion of the work. 1 Department of Mathematics, Southern Utah University, Cedar City, 84720, UT, USA.
    [Show full text]
  • Infinite Words As Determined by Languages of Their Prefixes
    INFINITEWORDSASDETERMINEDBY LANGUAGESOFTHEIRPREFIXES tim smith A dissertation submitted in partial fulfillment of the requirements for the degree of Ph.D. in computer science College of Computer and Information Science Northeastern University Boston, MA, USA April 2015 ABSTRACT We explore a notion of complexity for infinite words relating them to languages of their prefixes. An infinite language L determines an infinite word α if every string in L is a prefix of α. If L is regular, it is known that α must be ultimately periodic; conversely, every ul- timately periodic word is determined by some regular language. In this dissertation, we investigate other classes of languages and infi- nite words to see what connections can be made among them within this framework. We make particular use of pumping lemmas as a tool for studying these classes and their relationship to infinite words. • First, we investigate infinite words determined by various types of automata. We consider finite automata, pushdown automata, a generalization of pushdown automata called stack automata, and multihead finite automata, and relate these classes to ultimately periodic words and to a class of infinite words which we call mul- tilinear. • Second, we investigate infinite words determined by the parallel rewriting systems known as L systems. We show that certain infi- nite L systems necessarily have infinite subsets in other L systems, and use these relationships to categorize the infinite words deter- mined by a hierarchy of these systems, showing that it collapses to just three distinct classes of infinite words. • Third, we investigate infinite words determined by indexed gram- mars, a generalization of context-free grammars in which nonter- minals are augmented with stacks which can be pushed, popped, and copied to other nonterminals at the derivation proceeds.
    [Show full text]
  • Indexed Counter Languages Informatique Théorique Et Applications, Tome 26, No 1 (1992), P
    INFORMATIQUE THÉORIQUE ET APPLICATIONS J. DUSKE M. MIDDENDORF R. PARCHMANN Indexed counter languages Informatique théorique et applications, tome 26, no 1 (1992), p. 93-113 <http://www.numdam.org/item?id=ITA_1992__26_1_93_0> © AFCET, 1992, tous droits réservés. L’accès aux archives de la revue « Informatique théorique et applications » im- plique l’accord avec les conditions générales d’utilisation (http://www.numdam. org/conditions). Toute utilisation commerciale ou impression systématique est constitutive d’une infraction pénale. Toute copie ou impression de ce fichier doit contenir la présente mention de copyright. Article numérisé dans le cadre du programme Numérisation de documents anciens mathématiques http://www.numdam.org/ Informatique théorique et Applications/Theoretical Informaties and Applications (vol. 26, n° 1, 1992, p. 93 à 113) INDEXED COUNTER LANGUAGES (*) by J. DUSKE (x), M. MIDDENDORF (*) and R. PARCHMANN (*) Communicated by J. BERSTEL Abstract. — Starting with the characterization of context-free counter languages by rightlinear indexed grammars with one index, indexed counter grammars are introduced and investigated. The family of indexed counter languages is a full AFL properly contained in the family of indexed languages and incomparable with the full trio oflinear indexed languages. Furthermore by modifying the dérivation mode, a characterization of type-0 languages by indexed counter grammars is given. Résumé. - Après une caractêrisation des langages algébriques à compteurs par des grammaires indexées linéaires droites d'index 1, on introduit et étudie les grammaires indexées à compteurs. La famille des langages indexés à compteurs est une AFL proprement contenue dans la famille des langages indexés, et incomparable au cône rationnel des langages indexés linéaires.
    [Show full text]
  • Appendix A: Hierarchy of Grammar Formalisms
    Appendix A: Hierarchy of Grammar Formalisms The following figure recalls the language hierarchy that we have developed in the course of the book. '' $$ '' $$ '' $$ ' ' $ $ CFG, 1-MCFG, PDA n n L1 = {a b | n ≥ 0} & % TAG, LIG, CCG, tree-local MCTAG, EPDA n n n ∗ L2 = {a b c | n ≥ 0}, L3 = {ww | w ∈{a, b} } & % 2-LCFRS, 2-MCFG, simple 2-RCG & % 3-LCFRS, 3-MCFG, simple 3-RCG n n n n n ∗ L4 = {a1 a2 a3 a4 a5 | n ≥ 0}, L5 = {www | w ∈{a, b} } & % ... LCFRS, MCFG, simple RCG, MG, set-local MCTAG, finite-copying LFG & % Thread Automata (TA) & % PMCFG 2n L6 = {a | n ≥ 0} & % RCG, simple LMG (= PTIME) & % mildly context-sensitive L. Kallmeyer, Parsing Beyond Context-Free Grammars, Cognitive Technologies, DOI 10.1007/978-3-642-14846-0, c Springer-Verlag Berlin Heidelberg 2010 216 Appendix A For each class the different formalisms and automata that generate/accept exactly the string languages contained in this class are listed. Furthermore, examples of typical languages for this class are added, i.e., of languages that belong to this class while not belonging to the next smaller class in our hier- archy. The inclusions are all proper inclusions, except for the relation between LCFRS and Thread Automata (TA). Here, we do not know whether the in- clusion is a proper one. It is possible that both devices yield the same class of languages. Appendix B: List of Acronyms The following table lists all acronyms that occur in this book. (2,2)-BRCG Binary bottom-up non-erasing RCG with at most two vari- ables per left-hand side argument 2-SA Two-Stack Automaton
    [Show full text]
  • A Hierarchy of Mildly Context-Sensitive Dependency Grammars Anssi Yli-Jyra¨ and Matti Nykanen¨
    11 A Hierarchy of Mildly Context-Sensitive Dependency Grammars Anssi Yli-Jyra¨ and Matti Nykanen¨ Department of Modern Languages, University of Helsinki and School of Computing, University of Eastern Finland Email: Anssi.Yli-Jyra@helsinki.fi, Matti.Nykanen@uef.fi ABSTRACT. We generalize Link Grammars with crossing dependencies and pro- jective Dependency Grammars with a gradient projectivity condition and mul- tiple heads. The generalizations rely on right linear grammars that encode the links and dependencies via index storage instructions. The underlying machinery and the stepwise refinements of the definitions connect the generalized gram- mars to Linear Indexed Grammars and the mildly context sensitive hierarchy of Wartena’s right linear grammars that use composite index storages. The gradi- ent projectivity measure – the non-projectivty depth – relies on a normalized storage type whose runs represent particular multiplanar decompositions for de- pendency graphs. This decomposition does not aim at minimal graph coloring, but focuses, instead, to rigidly necessary color changes in right linear processing. Thanks to this multiplanar decomposition, every non-projectivity depth k charac- terizes strictly larger sets of acyclic dependency graphs and trees than the planar projective dependency graphs and trees. 11.1 Introduction In this paper, we generalize Link Grammars (Sleator and Temper- ley 1993) and projective Dependency Grammars (Hays 1964, Gaifman 1965) with crossing dependencies and then characterize the generaliza- tions using formal language theory. In terms of Dependency Grammars (DGs) (Tesni`ere1959), word- order is distinct from the dependency tree that analyzes the struc- ture of the sentence. However, Hays (1964) and Gaifman (1965) have Proceedings of Formal Grammar 2004.
    [Show full text]