Chart Parsing and Constraint Programming

Total Page:16

File Type:pdf, Size:1020Kb

Chart Parsing and Constraint Programming Chart Parsing and Constraint Programming Frank Morawietz Seminar f¨ur Sprachwissenschaft Universit¨at T¨ubingen Wilhelmstr. 113 72074 T¨ubingen, Germany [email protected] Abstract yet another language. The approach allows for a rapid In this paper, parsing-as-deduction and constraint pro- and very flexible but at the same time uniform method gramming are brought together to outline a procedure for of implementation of all kinds of parsing algorithms (for the specification of constraint-based chart parsers. Fol- constraint-based theories). The goal is not necessarily to lowing the proposal in Shieber et al. (1995), we show build the fastest parser, but rather to build – for an ar- how to directly realize the inference rules for deductive bitrary algorithm – a parser fast and perspicuously. For parsers as Constraint Handling Rules (Fr¨uhwirth, 1998) example, the advantage of our approach compared to the by viewing the items of a chart parser as constraints and one proposed in Shieber et al. (1995) is that we do not the constraint base as a chart. This allows the direct use have to design a special deduction engine and we do not of constraint resolution to parse sentences. have to handle chart and agenda explicitly. Furthermore, the process can be used in any constraint-based formal- 1 Introduction ism which allows for constraint propagation and there- fore can be easily integrated into existing applications. The parsing-as-deduction approach proposed in Pereira The paper proceeds by reviewing the parsing-as- and Warren (1983) and extended in Shieber et al. (1995) deduction approach and a particular way of imple- and the parsing schemata defined in Sikkel (1997) are menting constraint systems, Constraint Handling Rules well established parsing paradigms in computational lin- (CHR) as presented in Fr¨uhwirth (1998). Then it shows guistics. Their main strengths are their flexibility and how to implement several parsing algorithms very natu- the level of abstraction concerning control information rally with constraint propagation rules before concluding inherent in parsing algorithms. Furthermore, they are with an outlook on how to extend the technique to more easily extensible to more complex formalisms, e.g., aug- advanced applications. mented phrase structure rules or the ID/LP format. Constraint Programming (CP) has been used in com- 1.1 Parsing as Deduction putational linguistics in several areas, for example in Although I assume some familiarity with parsing-as- (typed) feature-based systems (Smolka, 1995), or condi- deduction, I will recall some basic definitions for con- tional constraints (Matiasek, 1994), or advanced compi- venience. The notations and the three basic algorithms lation techniques (G¨otz and Meurers, 1997) or special- are directly taken from Shieber et al. (1995). ized constraint solvers (Manandhar, 1994). But none As usual, strings w result from concatenation of sym- Σ Σ£ of these approaches uses constraint programming tech- bols from some alphabet set , i.e., w ¾ . We refer niques to implement standard chart parsing algorithms to the decomposition of such a string into its alphabet directly in a constraint system. symbols with indices. We fix this notation using w = ; ¾ Æ In this paper, I will bring these two paradigms to- w1 ::: wn. Further notational conventions are: i j , n ; ;::: gether by showing how to implement algorithms from for the length of the string to be parsed, A ; B C for ; ;::: the parsing-as-deduction scheme by viewing the parsing arbitrary formulas or nonterminals, a ; b c for termi- β ; γ;::: process as constraint propagation. nals, ε for the empty string and α; for strings of The core idea is that the items of a conventional chart terminals and nonterminals. Formulas used in parsing parser are constraints on labeled links between the words will also be called items or edges. A grammatical deduc- and positions of an input string. Then the inference rules tion system or, in Sikkel’s terminology a parsing schema, allow for the deduction of new constraints, again labeled is defined as a set of deduction schemes and a set of ax- and spanning parts of the input string, via constraint ioms. These are given with the help of formula schemata propagation. The resulting constraint store represents the which contain (syntactic) meta-variables which are in- chart which can be accessed to determine whether the stantiated with concrete terms on application of the rules. parse was successful or to reconstruct a parse tree. A deduction scheme R has the general form While this may seem a trivial observation, it is not A1 ::: An ::: ; i h side conditions on A A C just another way of implementing deductive parsing in C 1 n Table 1: Parsing algorithms as Grammatical Deduction Systems Bottom-Up Top-Down Earley ; α ¯ ] [ ; ¯ β ] [ ; ; ; α ¯ β ] Items [ j j i j A ¼ ; ¯ ] [ ; ¯ ] [ ; ; ; ¯ ] Axiom [0 0 S 0 0 S S ¼ ; ¯ ] [ ; ¯ ] [ ; ; ; ¯ ] Goal [n S n 0 n S S [ ; ¯ β ] [ ; ; ; α ¯ β ] [ ; α ¯ ] · j j w · i j A w Scan j 1 j 1 [ · ; α ¯ ] [ · ; ¯ β ] [ ; · ; ; α ¯ β ] · j 1 w j ·1 j 1 i j 1 A w j 1 ; ¯ β ] [ ; ; ; α ¯ β ] [ j B i j A B ! γi h ! γi Predict hB B ; ¯ γβ ] [ ; ; ; ¯ γ] [ j j j B ; αγ ¯ ] [ ; ; ; α ¯ β ] [ ; ; ; γ ¯ ] [ j i k A B k j B ! γi Complete hB ; α ¯ ] [ ; ; ; α ¯ β ] [ j B i j A B ! ! where the Ai and C are formula schemata. The Ai are S NP VP V hit ! j ! called antecedents and C the consequence. Note that de- VP VNP VNPPP PN John duction schemes may refer to string positions, i.e., the ! ! j PP PNP N dog stick indices within the input string, in their side conditions. ! j ! NP PN Det N1 P with Application of these schemata and derivations of formu- ! j ! las are then defined as in the Shieber et al. article. In- N1 N N1 PP Det the tuitively, parsing uses the deductive rules – if their an- tecedents and the side conditions are met – to infer new Figure 1: Example Grammar: PP-attachment items from the axioms and already generated items until no new ones can be derived. The parse was successful if maintain a constraint base or store which is continually a goal item was derived. monitored for possible rule applications, i.e., whether Therefore, all the parsing systems used in this paper there is enough information present to successfully use a are defined by specifying a class of items, a set of ax- rule to simplify constraints or to derive new constraints. ioms, a set of inference rules and a subset of the items, Whereas usually one deals with a fixed constraint domain the goals. For better readability, I follow Shieber et al. in and a specialized solver, CHR is an extension of the Pro- using the familiar dotted items for the presentation. The log language which allows for the specification of user- three classical example algorithms we will use to illus- defined constraints and arbitrary solvers. The strength of trate our technique are given in Tab. 1. I assume famil- the CHR approach lies in the fact that it allows for mul- iarity with these algorithms. tiple (conjunctively interpreted) heads in rules, that it is Unless specified differently, we assume that we are flexible and that it is tightly and transparently integrated into the Prolog engine. h ; Σ; ; i given a context-free grammar G = N S P with non- terminals N,terminalsΣ, start symbol S and set of pro- In CHR constraints are just distinguished sets of ductions P. For Earley’s algorithm we also assume a new (atomic) formulas. CHR allow the definition of rule sets start symbol S ¼ which is not in N. Each production is of for constraint solving with three types of rules: Firstly <=> £ simplification rules ( ) which replace a number of ! α ¾ ; α ¾ ´ [ Σµ the form A with A N N . For examples I will use the simple PP-attachment grammar G given in constraints in the store with new constraints; secondly > Fig. 1 with the obvious sets of nonterminals and termi- propagation rules ( == ) which add new constraints nals, the start symbol S and productions P. Itisleftto to the store in case a number of constraints is already => the reader to calculate example derivations for the three present; and thirdly “simpagation” rules ( < in com- algorithms for a sentence such as John hit the dog with bination with a Ò in the head of the rule) which replace the stick. only those constraints with new ones which are to the right of the backslash. Rules can have guards. A guard 1.2 Constraint Handling Rules (separated from the rest of the body by a j ) is a condition There are several constraint programming environments which has to be met before the rule can be applied. available. The most recent and maybe the most flexible is We cannot go into the details of the formal seman- the Constraint Handling Rules (CHR) package included tics of CHR here. The interested reader is referred to in SICStus Prolog (Fr¨uhwirth, 1998). These systems Fr¨uhwirth (1998). Since I will refer back to it let us just note that logically, simplification rules are equivalences parse(InList):- and propagation rules are implications if their guard is axiom, satisfied. Simpagation rules are special cases of simpli- post_const(InList, 0, Length), fication rules. Soundness and completeness results for report(Length). CHR are available (Abdennadher et al., 1996 Abdennad- her, 1998). post_const([], Len, Len). post_const([Word|Str], InLen, Len):- 2 Parsing as Constraint Propagation findall(Cat, lex(Word,Cat), Cats), post_words(Cats, InLen, Word), The basic observation which turns parsing-as-deduction NewLen is InLen + 1, into constraint propagation is simple: items of a chart post_const(Str, NewLen, Len).
Recommended publications
  • Journal of Computer Science and Engineering Parsing
    IJRDO - Journal of Computer Science and Engineering ISSN: 2456-1843 JOURNAL OF COMPUTER SCIENCE AND ENGINEERING PARSING TECHNIQUES Ojesvi Bhardwaj Abstract: ‘Parsing’ is the term used to describe the process of automatically building syntactic analyses of a sentence in terms of a given grammar and lexicon. The resulting syntactic analyses may be used as input to a process of semantic interpretation, (or perhaps phonological interpretation, where aspects of this, like prosody, are sensitive to syntactic structure). Occasionally, ‘parsing’ is also used to include both syntactic and semantic analysis. We use it in the more conservative sense here, however. In most contemporary grammatical formalisms, the output of parsing is something logically equivalent to a tree, displaying dominance and precedence relations between constituents of a sentence, perhaps with further annotations in the form of attribute-value equations (‘features’) capturing other aspects of linguistic description. However, there are many different possible linguistic formalisms, and many ways of representing each of them, and hence many different ways of representing the results of parsing. We shall assume here a simple tree representation, and an underlying context-free grammatical (CFG) formalism. *Student, CSE, Dronachraya Collage of Engineering, Gurgaon Volume-1 | Issue-12 | December, 2015 | Paper-3 29 - IJRDO - Journal of Computer Science and Engineering ISSN: 2456-1843 1. INTRODUCTION Parsing or syntactic analysis is the process of analyzing a string of symbols, either in natural language or in computer languages, according to the rules of a formal grammar. The term parsing comes from Latin pars (ōrātiōnis), meaning part (of speech). The term has slightly different meanings in different branches of linguistics and computer science.
    [Show full text]
  • An Efficient Implementation of the Head-Corner Parser
    An Efficient Implementation of the Head-Corner Parser Gertjan van Noord" Rijksuniversiteit Groningen This paper describes an efficient and robust implementation of a bidirectional, head-driven parser for constraint-based grammars. This parser is developed for the OVIS system: a Dutch spoken dialogue system in which information about public transport can be obtained by telephone. After a review of the motivation for head-driven parsing strategies, and head-corner parsing in particular, a nondeterministic version of the head-corner parser is presented. A memorization technique is applied to obtain a fast parser. A goal-weakening technique is introduced, which greatly improves average case efficiency, both in terms of speed and space requirements. I argue in favor of such a memorization strategy with goal-weakening in comparison with ordinary chart parsers because such a strategy can be applied selectively and therefore enormously reduces the space requirements of the parser, while no practical loss in time-efficiency is observed. On the contrary, experiments are described in which head-corner and left-corner parsers imple- mented with selective memorization and goal weakening outperform "standard" chart parsers. The experiments include the grammar of the OV/S system and the Alvey NL Tools grammar. Head-corner parsing is a mix of bottom-up and top-down processing. Certain approaches to robust parsing require purely bottom-up processing. Therefore, it seems that head-corner parsing is unsuitable for such robust parsing techniques. However, it is shown how underspecification (which arises very naturally in a logic programming environment) can be used in the head-corner parser to allow such robust parsing techniques.
    [Show full text]
  • Generalized Probabilistic LR Parsing of Natural Language (Corpora) with Unification-Based Grammars
    Generalized Probabilistic LR Parsing of Natural Language (Corpora) with Unification-Based Grammars Ted Briscoe* John Carroll* University of Cambridge University of Cambridge We describe work toward the construction of a very wide-coverage probabilistic parsing system for natural language (NL), based on LR parsing techniques. The system is intended to rank the large number of syntactic analyses produced by NL grammars according to the frequency of occurrence of the individual rules deployed in each analysis. We discuss a fully automatic procedure for constructing an LR parse table from a unification-based grammar formalism, and consider the suitability of alternative LALR(1) parse table construction methods for large grammars. The parse table is used as the basis for two parsers; a user-driven interactive system that provides a computationally tractable and labor-efficient method of supervised training of the statistical information required to drive the probabilistic parser. The latter is constructed by associating probabilities with the LR parse table directly. This technique is superior to parsers based on probabilistic lexical tagging or probabilistic context-free grammar because it allows for a more context-dependent probabilistic language model, as well as use of a more linguistically adequate grammar formalism. We compare the performance of an optimized variant of Tomita's (1987) generalized LR parsing algorithm to an (efficiently indexed and optimized) chart parser. We report promising results of a pilot study training on 150 noun definitions from the Longman Dictionary of Contemporary English (LDOCE) and retesting on these plus a further 55 definitions. Finally, we discuss limitations of the current system and possible extensions to deal with lexical (syntactic and semantic)frequency of occurrence.
    [Show full text]
  • CS 375, Compilers: Class Notes Gordon S. Novak Jr. Department Of
    CS 375, Compilers: Class Notes Gordon S. Novak Jr. Department of Computer Sciences University of Texas at Austin [email protected] http://www.cs.utexas.edu/users/novak Copyright c Gordon S. Novak Jr.1 1A few slides reproduce figures from Aho, Lam, Sethi, and Ullman, Compilers: Principles, Techniques, and Tools, Addison-Wesley; these have footnote credits. 1 I wish to preach not the doctrine of ignoble ease, but the doctrine of the strenuous life. { Theodore Roosevelt Innovation requires Austin, Texas. We need faster chips and great compilers. Both those things are from Austin. { Guy Kawasaki 2 Course Topics • Introduction • Lexical Analysis: characters ! words lexer { Regular grammars { Hand-written lexical analyzer { Number conversion { Regular expressions { LEX • Syntax Analysis: words ! sentences parser { Context-free grammars { Operator precedence { Recursive descent parsing { Shift-reduce parsing, YACC { Intermediate code { Symbol tables • Code Generation { Code generation from trees { Register assignment { Array references { Subroutine calls • Optimization { Constant folding, partial evaluation, Data flow analysis • Object-oriented programming 3 Pascal Test Program program graph1(output); { Jensen & Wirth 4.9 } const d = 0.0625; {1/16, 16 lines for [x,x+1]} s = 32; {32 character widths for [y,y+1]} h = 34; {character position of x-axis} c = 6.28318; {2*pi} lim = 32; var x,y : real; i,n : integer; begin for i := 0 to lim do begin x := d*i; y := exp(-x)*sin(c*x); n := round(s*y) + h; repeat write(' '); n := n-1 until n=0; writeln('*') end end. * * * * * * * * * * * * * * * * * * * * * * * * * * 4 Introduction • What a compiler does; why we need compilers • Parts of a compiler and what they do • Data flow between the parts 5 Machine Language A computer is basically a very fast pocket calculator attached to a large memory.
    [Show full text]
  • LATE Ain't Earley: a Faster Parallel Earley Parser
    LATE Ain’T Earley: A Faster Parallel Earley Parser Peter Ahrens John Feser Joseph Hui [email protected] [email protected] [email protected] July 18, 2018 Abstract We present the LATE algorithm, an asynchronous variant of the Earley algorithm for pars- ing context-free grammars. The Earley algorithm is naturally task-based, but is difficult to parallelize because of dependencies between the tasks. We present the LATE algorithm, which uses additional data structures to maintain information about the state of the parse so that work items may be processed in any order. This property allows the LATE algorithm to be sped up using task parallelism. We show that the LATE algorithm can achieve a 120x speedup over the Earley algorithm on a natural language task. 1 Introduction Improvements in the efficiency of parsers for context-free grammars (CFGs) have the potential to speed up applications in software development, computational linguistics, and human-computer interaction. The Earley parser has an asymptotic complexity that scales with the complexity of the CFG, a unique, desirable trait among parsers for arbitrary CFGs. However, while the more commonly used Cocke-Younger-Kasami (CYK) [2, 5, 12] parser has been successfully parallelized [1, 7], the Earley algorithm has seen relatively few attempts at parallelization. Our research objectives were to understand when there exists parallelism in the Earley algorithm, and to explore methods for exploiting this parallelism. We first tried to naively parallelize the Earley algorithm by processing the Earley items in each Earley set in parallel. We found that this approach does not produce any speedup, because the dependencies between Earley items force much of the work to be performed sequentially.
    [Show full text]
  • NLTK Parsing Demos
    NLTK Parsing Demos Adrian Brasoveanu∗ March 3, 2014 Contents 1 Recursive descent parsing (top-down, depth-first)1 2 Shift-reduce (bottom-up)4 3 Left-corner parser: top-down with bottom-up filtering5 4 General Background: Memoization6 5 Chart parsing 11 6 Bottom-Up Chart Parsing 15 7 Top-down Chart Parsing 21 8 The Earley Algorithm 25 9 Back to Left-Corner chart parsing 30 1 Recursive descent parsing (top-down, depth-first) [py1] >>> import nltk, re, pprint >>> from __future__ import division [py2] >>> grammar1= nltk.parse_cfg(""" ... S -> NP VP ... VP -> V NP | V NP PP ... PP -> P NP ... V ->"saw"|"ate"|"walked" ∗Based on the NLTK book (Bird et al. 2009) and created with the PythonTeX package (Poore 2013). 1 ... NP ->"John"|"Mary"|"Bob" | Det N | Det N PP ... Det ->"a"|"an"|"the"|"my" ... N ->"man"|"dog"|"cat"|"telescope"|"park" ... P ->"in"|"on"|"by"|"with" ... """) [py3] >>> rd_parser= nltk.RecursiveDescentParser(grammar1) >>> sent=’Mary saw a dog’.split() >>> sent [’Mary’, ’saw’, ’a’, ’dog’] >>> trees= rd_parser.nbest_parse(sent) >>> for tree in trees: ... print tree,"\n\n" ... (S (NP Mary) (VP (V saw) (NP (Det a) (N dog)))) [py4] >>> rd_parser= nltk.RecursiveDescentParser(grammar1, trace=2) >>> rd_parser.nbest_parse(sent) Parsing ’Mary saw a dog’ [*S] E [ * NP VP ] E [ * ’John’ VP ] E [ * ’Mary’ VP ] M [ ’Mary’ * VP ] E [ ’Mary’ * V NP ] E [ ’Mary’ * ’saw’ NP ] M [ ’Mary’ ’saw’ * NP ] E [ ’Mary’ ’saw’ * ’John’ ] E [ ’Mary’ ’saw’ * ’Mary’ ] E [ ’Mary’ ’saw’ * ’Bob’ ] E [ ’Mary’ ’saw’ * Det N ] E [ ’Mary’ ’saw’ * ’a’ N ] M [ ’Mary’ ’saw’
    [Show full text]
  • Parsing 1. Grammars and Parsing 2. Top-Down and Bottom-Up Parsing 3
    Syntax Parsing syntax: from the Greek syntaxis, meaning “setting out together or arrangement.” 1. Grammars and parsing Refers to the way words are arranged together. 2. Top-down and bottom-up parsing Why worry about syntax? 3. Chart parsers • The boy ate the frog. 4. Bottom-up chart parsing • The frog was eaten by the boy. 5. The Earley Algorithm • The frog that the boy ate died. • The boy whom the frog was eaten by died. Slide CS474–1 Slide CS474–2 Grammars and Parsing Need a grammar: a formal specification of the structures allowable in Syntactic Analysis the language. Key ideas: Need a parser: algorithm for assigning syntactic structure to an input • constituency: groups of words may behave as a single unit or phrase sentence. • grammatical relations: refer to the subject, object, indirect Sentence Parse Tree object, etc. Beavis ate the cat. S • subcategorization and dependencies: refer to certain kinds of relations between words and phrases, e.g. want can be followed by an NP VP infinitive, but find and work cannot. NAME V NP All can be modeled by various kinds of grammars that are based on ART N context-free grammars. Beavis ate the cat Slide CS474–3 Slide CS474–4 CFG example CFG’s are also called phrase-structure grammars. CFG’s Equivalent to Backus-Naur Form (BNF). A context free grammar consists of: 1. S → NP VP 5. NAME → Beavis 1. a set of non-terminal symbols N 2. VP → V NP 6. V → ate 2. a set of terminal symbols Σ (disjoint from N) 3.
    [Show full text]
  • Chart Parsing and Probabilistic Parsing
    Chapter 8 Chart Parsing and Probabilistic Parsing 8.1 Introduction Chapter 7 started with an introduction to constituent structure in English, showing how words in a sentence group together in predictable ways. We showed how to describe this structure using syntactic tree diagrams, and observed that it is sometimes desirable to assign more than one such tree to a given string. In this case, we said that the string was structurally ambiguous; and example was old men and women. Treebanks are language resources in which the syntactic structure of a corpus of sentences has been annotated, usually by hand. However, we would also like to be able to produce trees algorithmically. A context-free phrase structure grammar (CFG) is a formal model for describing whether a given string can be assigned a particular constituent structure. Given a set of syntactic categories, the CFG uses a set of productions to say how a phrase of some category A can be analyzed into a sequence of smaller parts ±1 ... ±n. But a grammar is a static description of a set of strings; it does not tell us what sequence of steps we need to take to build a constituent structure for a string. For this, we need to use a parsing algorithm. We presented two such algorithms: Top-Down Recursive Descent (7.5.1) and Bottom-Up Shift-Reduce (7.5.2). As we pointed out, both parsing approaches suffer from important shortcomings. The Recursive Descent parser cannot handle left-recursive productions (e.g., productions such as NP → NP PP), and blindly expands categories top-down without checking whether they are compatible with the input string.
    [Show full text]
  • Chart Parsing: the CYK Algorithm Complexity
    Problems with Parsing as Search Problems with Parsing as Search The CYK Algorithm The CYK Algorithm 1 Problems with Parsing as Search Grammar Restructuring Chart Parsing: the CYK Algorithm Complexity Informatics 2A: Lecture 16 2 The CYK Algorithm Parsing as Dynamic Programming The CYK Algorithm Bonnie Webber Visualizing the Chart School of Informatics Properties of the Algorithm University of Edinburgh [email protected] Reading: 30 October 2009 J&M (2nd ed), ch. 13, Section 13.3 – 13.4 NLTK Book, ch. 8 (Analyzing Sentence Structure), Section 8.4 Informatics 2A: Lecture 16 Chart Parsing: the CYK Algorithm 1 Informatics 2A: Lecture 16 Chart Parsing: the CYK Algorithm 2 Problems with Parsing as Search Grammar Restructuring Problems with Parsing as Search Grammar Restructuring The CYK Algorithm Complexity The CYK Algorithm Complexity Grammar Restructuring Left Recursion But grammars for natural human languages should be revealing, allowing strings to be associated with meaning (semantics) in a Deterministic parsing (Lectures 14 and 15) aims to address a systematic way. Re-structuring the grammar may destroy this. limited amount of local ambiguity – the problem of not being able to decide uniquely which grammar rule to use next in a Example: (Indirectly) left-recursive rules are needed in English. left-to-right analysis of the input string, even if the string is not NP → DET N globally ambiguous. NP → NPR By re-structuring the grammar, the parser can make a unique DET → NP ’s decision, based on a limited amount of look-ahead. These rules generate NPs with possesive modifiers such as: Recursive Descent parsing (Lecture 13) demands grammar restructuring, in order to eliminate left-recursive rules that can get John’s sister it into a hopeless loop.
    [Show full text]
  • Compiler Construction
    Compiler construction PDF generated using the open source mwlib toolkit. See http://code.pediapress.com/ for more information. PDF generated at: Sat, 10 Dec 2011 02:23:02 UTC Contents Articles Introduction 1 Compiler construction 1 Compiler 2 Interpreter 10 History of compiler writing 14 Lexical analysis 22 Lexical analysis 22 Regular expression 26 Regular expression examples 37 Finite-state machine 41 Preprocessor 51 Syntactic analysis 54 Parsing 54 Lookahead 58 Symbol table 61 Abstract syntax 63 Abstract syntax tree 64 Context-free grammar 65 Terminal and nonterminal symbols 77 Left recursion 79 Backus–Naur Form 83 Extended Backus–Naur Form 86 TBNF 91 Top-down parsing 91 Recursive descent parser 93 Tail recursive parser 98 Parsing expression grammar 100 LL parser 106 LR parser 114 Parsing table 123 Simple LR parser 125 Canonical LR parser 127 GLR parser 129 LALR parser 130 Recursive ascent parser 133 Parser combinator 140 Bottom-up parsing 143 Chomsky normal form 148 CYK algorithm 150 Simple precedence grammar 153 Simple precedence parser 154 Operator-precedence grammar 156 Operator-precedence parser 159 Shunting-yard algorithm 163 Chart parser 173 Earley parser 174 The lexer hack 178 Scannerless parsing 180 Semantic analysis 182 Attribute grammar 182 L-attributed grammar 184 LR-attributed grammar 185 S-attributed grammar 185 ECLR-attributed grammar 186 Intermediate language 186 Control flow graph 188 Basic block 190 Call graph 192 Data-flow analysis 195 Use-define chain 201 Live variable analysis 204 Reaching definition 206 Three address
    [Show full text]
  • A Modified Earley Parser for Huge Natural Language Grammars 2 Earley Parser
    A Modified Earley Parser for Huge Natural Language Grammars Sinan Polat, Merve Selcuk-Simsek, and Ilyas Cicekli Department of Computer Engineering Hacettepe University Ankara, Turkey [email protected] fmerveselcuk, [email protected] Abstract. For almost a half century Earley parser has been used in the parsing of context-free grammars and it is considered as a touch-stone algorithm in the history of parsing algorithms. On the other hand, it is also known for being expensive from its time requirement and memory usage perspectives. For huge context-free grammars, its performance is not good since its time complexity also depends on the number of rules in the grammar. The time complexity of the original Earley parser is O(R2N 3) where N is the string length, and R is the number of rules. In this paper, we aim to improve time and memory usage performances of Earley parser for grammars with a large number of rules. In our approach, we prefer radix tree representation for rules instead of list representation as in original Earley parser. We have tested our algorithm using different number of rule sets up to 200,000 which are all learned by an example- based machine translation system. According to our evaluation results, our modified parser has a time bound of O(log(R)N 3), and it has 20% less memory usage regarding the original Earley parser. Keywords: earley parser, computational linguistics, natural language processing 1 Introduction Grammar parsers first revealed with programming languages. Until Earley's algorithm, the previous algorithms had been able only to parse unambiguous grammars, or subsets of context-free grammars.
    [Show full text]
  • A Hints and Solutions to Selected Problems
    A Hints and Solutions to Selected Problems Answer to Problem 2.1: The description is not really finite: it depends on all descriptions in the list, of which there are infinitely many. Answer to Problem 2.2: Any function that maps the integers onto unique integers will do (an injective function), for example nn or the n-th prime. Answer to Problem 2.3: Hint: use a special (non-terminal) marker for each block the tur- tle is located to the east of its starting point. Answer to Problem 2.4: The terminal production cannot contain S, and the only way to get rid of it is to apply S--->aSQ k ≥ 0 times followed by S--->abc. This yields a sentential form akabcQk. The part bcQk cannot produce any as, so all the as in the result come from aka,so all strings an··· can be produced, for n = k+1 ≥ 1. We now need to prove that bcQk produces exactly bncn. The only way to get rid of the Q is through bQc--->bbcc, which requires the Q to be adjacent to the b. The only way to get it there is through repeated application of cQ--->Qc. This does not affect the number of each token. After j application of bQc--->bbcc we have bb jc jcQk − j, and after k applications we have bbkckc, which equals bncn. Answer to Problem 2.5: Assume the alphabet is {(, [}. First create a nesting string of (, ), [,and], with a marker in the center and another marker at the end: ([[...[([X])]...]])Z.
    [Show full text]