Using Decision Trees to Select the Grammatical Relation of a Noun Phrase

Total Page:16

File Type:pdf, Size:1020Kb

Using Decision Trees to Select the Grammatical Relation of a Noun Phrase Using decision trees to select the gran natical relation of a noun phrase Simon CORSTON-OLIVER Microsoft Research One Microsoft Way Redmond WA 98052, USA [email protected] Abstract For example, transitive clauses tend not to contain We present a machine-learning approach to lexical NPs in both subject and object positions and modeling the distribution of noun phrases subjects of transitives tend not to be lexical NPs nor (NPs) within clauses with respect to a fine- to be discourse-new. grained taxonomy of grammatical relations. We Unfortunately, the models used in PAS have demonstrate that a cluster of superficial involved only simple chi-squared tests to identify linguistic features can function as a proxy for statistically significant patterns in the distribution of more abstract discourse features that are not NPs with respect to pairs of features (e.g. part of observable using state-of-the-art natural speech and grammatical relation). A further problem language processing. The models constructed from the point of view of computational discourse for actual texts can be used to select among analysis is that many of the features used in empirical alternative linguistic expressions of the same studies are not observable in texts using state-of-the propositional content when generating art natural language processing. Such non-observable discourse. features include animacy, the information status of a referent, and the identification of the gender of a 1. Introduction referent based on world knowledge. Natural language generation involves a number of In the present study, we treat the task of processes ranging from planning the content to be determining the appropriate distribution of mentions expressed through making encoding decisions in text as a machine learning classification problem: involving syntax, the lexicon and morphology. The what is the probability that a mention will have a present study concerns decisions made about the form certain grammatical relation given a deh set of and distribution of each "mention" of a discourse linguistic features? In particular, how accurately can entity: should reference be made with a lexical NP, a we select appropriate grammatical relations using pronominal NP or a zero anaphor (i.e. an elided only superficial linguistic features? mention)? Should a given mention be expressed as the subject of its clause or in some other grammatical 2. Data relation? A total of 5,252 mentions were annotated from the If all works well, a natural language generation Encarta electronic encyclopedia and 4,937 mentions system may end up proposing a mmaber of possible from the Wall Street Journal (WSJ). Sentences were well-formed expressions of the same propositional parsed using the Microsoft English Grammar content. Although these possible formulations would (Heidorn 1999) to extract mentions and linguistic all be judged to be valid sentences of the target features. These analyses were then hand-corrected to language, it is not the ease that they are all equally eliminate noise in the training data caused by likely to occur. inaccurate parses, allowing us to determine the upper Research in the area of Preferred Argument bound on accuracy for the classification task if the Structure (Corston 1996, Du Bois 1987) has computational analysis were perfect. Zero anaphors established that in discourse in many languages, were annotated only when they occurred as subjects including English, NPs are distributed across of coordinated clauses. They have been excluded grammatical relations in statistically significant ways. 66 from the present study since they are invariably • [NounClass] We distinguish common nouns discourse-given subjects. versus proper names. Within proper names, 3. Features we distinguish the name of a place ("Geo") versus other proper names ("ProperName"). Nineteen linguistic features were annotated, along with information about the referent of each mention. • [Plural] The head of the mention is On the basis of the reference information we morphologically marked as plural. extracted the feature [InformationStatus], distinguishing "discourse-new" versus "discourse- • [POS] The part of speech of the head of the old". All mentions without a prior coreferential mention. mention in the text were classified as discourse-new, • [Prep] The governing preposition, if any. even if they would not traditionally be considered referential. [InformationStatus] is not directly • [RelC1] The mention is a child of a relative observable since it requires the analyst to make clause. decisions about the referent of a mention. • [TopLevel] The mention is not embedded In addition to the feature [InformafionStatus], the within another mention. following eighteen observable features were annotated. These are all features that we can • [Words] The total number of words in the reasonably expect syntactic parsers to extract with mention, discretized to the following values: sufficient accuracy today or in the near future. {0, 1, 2, 3, 4, 5, 6to10, 1 lto15, abovel5}. Gender ([Fern], [Mast]) was only annotated for • [ClausalStatus]: Does the mention occur in a common nouns whose default word sense is gendered main clause ("M"), complement clause ("C"), (e.g. "mother", "father"), for common nouns with or subordinate clause ("S")? specific morphology (e.g. with the -ess suffix) and for gender-marked proper names (e.g. "John", • [Coordinated] The mention is coordinated "Mary"). Gender was not marked for pronouns, to with at least one sibling. avoid difficult encoding decisions such as the use of • [Definite] The mention is marked with the genetic "he". ~ Gender was also not marked for cases definite article or a demonstrative pronoun. that would require world knowledge. The feature [GrRel] was given a much finer- [Fem] The mention is unambiguously grained analysis than is usual in computational feminine. linguistics. Studies in PAS have demonstrated the • [GrRel] The grammatical relation of the need to distinguish finer-grained categories than the mention (see below, this section). traditional grammatical relations of English grammar ("subject", "object" ere) in order to account for • [HasPossessive] Modified by a possessive distributional phenomena in discourse. For example, pronoun or a possessive NP with the elitic's subjects of intransitive verbs pattern with the direct ors'. objects of transitive verbs as being the preferred locus • [HasPP] Contains a postmodifying pre- for introducing new mentions. Subjects of transitives, positional phrase. however, are strongly dispreferred slots for the expression of new information. The use of fine- • [HasRelC1] Contains a postmodifying relative grained grammatical relations enables us to make clause. rather specific claims about the distribution of • [InQuotes] The mention occurs in quoted mentions. The taxonomy of fine-grained grammatical material. relations is given below in Figure 1. • [Lex] The specific inflected form of a pronoun, e.g. he, him. • [Mase] The mention is unambiguously 1 The feature [Lex] was sufficient for the decision tree tools masculine. to learn idiosyncratic uses of gendered pronouns. 67 Subject of transitive (S.r) Subject of copula Subject (Sc) Subject of intransitive Object of transitive Subject of intransitive (non-copula)(Si) ///L (PN) Grammatical Po so.Po PCP e I Relation ,//'t, NPCPP. I ~! adjective(PP,) } ., .. I ~ PP complementof J ' N°un (NA~=ve I [ verb (PPv) I Other (Oth) J Figure 1 The taxonomy of grammatical relations where 0 < k _< 1, and c is a constant such that p(S) 4. Decision trees sums to one. Note that smaller values of kappa cause For a set of annotated examples, we used decision- simpler structures to be favored. As kappa grows tree tools to construct the conditional probability of a closer to one (k = 1 corresponds to a uniform prior specific grammatical relation, given other features in over all possible tree structures), the learned decision the domain, z The decision trees are constructed using trees become more elaborate. Decision trees were a Bayesian learning approach that identifies tree built for k~ {0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, structures with high posterior probability (Chickefing et al. 1997). In particular, a candidate tree structure 0.95, 0.99, 0.999}. Having selected a decision tree, we use the (S) is evaluated against data (D) using Bayes' rule as posterior means of the parameters to specify a follows: probability distribution over the grammatical P(SID) = constant- p(DIS) • p(S) relations. To avoid overfitting, nodes containing fewer than fifty examples were not split during the For simplicity, we specify a prior distribution learning process. In building decision trees, 70% of over tree structures using a single parameter kappa the data was used for training and 30% for held-out (k). Assuming that N(S) probabilities are needed to evaluation. parameterize a tree with structure S, we use: The decision trees constructed can be rather complex, making them difficult to present visually. Figure 2 p(S) c. k = gives a simpler decision tree that predicts the grammatical relation of a mention for Enearta at 2 Comparison experiments were also done with Support Vector Machines (Platt 2000, Vapnik 1998) using a variety of kernel functions. The results obtained were indistinguishable from those reported here. 68 ,=:..,°IY" ~)=o.~ I ,- "-. I t~,te,n}=o.o~ ._.....2 °~'---. It~=:,=o.o~ p(sc)=O.09 I I p(na)=0.O~ ] I p(nay=u..26 J p(ppv)=O.~ I P(~) =0-07 1 y ~ -~, p(sc}=0.1~l P~PPW~.-3 p(st)=o,os ~ .;~..~Zo I~(st)=o.o3 I I p(ppa)=o.o5 ~(poss)=o.: p(srn)=O.Ol p(so)=o.o2 Figure 2 Decision tree for Enearta, at k=0.7 k=0.7. The tree was constructed using a subset of the 5. Evaluating decision trees morphological and syntactic features: [Coordinated], Decision trees were constructed and evaluated for [HasPP], [Lex], [NounClass], [Plural], [POS], [Prep], each corpus. We were particularly interested in the [RelC1], [TopLevel], [Words]. Grammatical relations accuracy of models built using only observable with only a residual probability are omitted for the features.
Recommended publications
  • Chapter 30 HPSG and Lexical Functional Grammar Stephen Wechsler the University of Texas Ash Asudeh University of Rochester & Carleton University
    Chapter 30 HPSG and Lexical Functional Grammar Stephen Wechsler The University of Texas Ash Asudeh University of Rochester & Carleton University This chapter compares two closely related grammatical frameworks, Head-Driven Phrase Structure Grammar (HPSG) and Lexical Functional Grammar (LFG). Among the similarities: both frameworks draw a lexicalist distinction between morphology and syntax, both associate certain words with lexical argument structures, both employ semantic theories based on underspecification, and both are fully explicit and computationally implemented. The two frameworks make available many of the same representational resources. Typical differences between the analyses proffered under the two frameworks can often be traced to concomitant differ- ences of emphasis in the design orientations of their founding formulations: while HPSG’s origins emphasized the formal representation of syntactic locality condi- tions, those of LFG emphasized the formal representation of functional equivalence classes across grammatical structures. Our comparison of the two theories includes a point by point syntactic comparison, after which we turn to an exposition ofGlue Semantics, a theory of semantic composition closely associated with LFG. 1 Introduction Head-Driven Phrase Structure Grammar is similar in many respects to its sister framework, Lexical Functional Grammar or LFG (Bresnan et al. 2016; Dalrymple et al. 2019). Both HPSG and LFG are lexicalist frameworks in the sense that they distinguish between the morphological system that creates words and the syn- tax proper that combines those fully inflected words into phrases and sentences. Stephen Wechsler & Ash Asudeh. 2021. HPSG and Lexical Functional Gram- mar. In Stefan Müller, Anne Abeillé, Robert D. Borsley & Jean- Pierre Koenig (eds.), Head-Driven Phrase Structure Grammar: The handbook.
    [Show full text]
  • Grammatical Relations Typology
    Draft of a chapter for The Oxford Handbook of Language Typology, ed. Jae Jung Song. Grammatical Relations Typology Balthasar Bickel University of Leipzig 1. Grammatical relations past and present Traditionally, the term ‘grammatical relation’ (GR) refers to the morphosyntactic pro- perties that relate an argument to a clause, as, for example, its subject or its object. Alternative terms are ‘syntactic function’ or ‘syntactic role’, and they highlight the fact that GRs are defined by the way in which arguments are integrated syntactically into a clause, i.e. by functioning as subject, object etc. Whatever terminology one prefers, what is crucial about the traditional notion of GRs is (a) that they are identi- fied by syntactic properties, and (b) that they relate an argument to the clause. This differentiates GRs from semantic roles (SRs), also known as thematic roles (θ-roles): SRs are semantic, not syntactic relations, and they hold between arguments and predicates (typically verbs), rather than between arguments and clauses. The difference between GRs and SRs is best visible in such contrasts as Sue killed the shark vs. Sue was killed by the shark: in both cases, the NP Sue is the subject of the clause. But in the active version, the referent of Sue is the agent of ‘kill’, while in the passive version, Sue is the patient of ‘kill.’ The syntactic properties that have traditionally been considered the key identifiers of GRs are the property of triggering verb agreement and the property of beeing assigned a specific case. In our example, Sue triggers third person singular agreement in the verb and this identifies the NP as the subject of the clause.
    [Show full text]
  • From Lexical to Functional Categories: New Foundations for the Study of Language Development
    From Lexical to Functional Categories: New Foundations for the Study of Language Development Cristina Dye (Newcastle University), Yarden Kedar (Beit Berl College) & Barbara Lust (Cornell University) 1. Introduction We will use the term Functional Categories (FCs) in this article in contrast to the term Lexical Categories (LCs) to capture a distinction between the nouns, verbs and adjectives (LCs) exemplified in (1) below versus the underlined determiners (the, an), the verb phrase markers (the auxiliary will and the morphological verb inflection -s), the preposition (with), and the complementizer (that) – all of which constitute FCs in this sentence. (1) The young girl will eat an ice cream that her friend shares with her. All natural languages appear to distinguish FCs and LCs (e.g., Abney, 1987; Chomsky, 1995), and FCs, as a fundamental part of syntax, are thought to be unique to human languages (Hauser, Barner, & O’Donnell, 2007, p. 111). In general, FCs are assumed to carry major grammatical weight in sentence construction, whereas the LCs are assumed to carry major semantic force. However, we must recognize the difficulty in making this distinction. For example, even FCs may carry semantic relevance, and some units appear to be mixed, that is, the small class of pronouns such as her in (1) above which involves limited semantic factors such as gender, and plays a functional role in determining phrase structure. In addition, FCs vary across languages in their range and nature. In spite of the fact that FCs are often viewed as representing a small, closed class word category, it is now recognized that “the number of functional elements in syntax is not easy to estimate..
    [Show full text]
  • Chapter 4: Syntactic Relations and Case Marking
    Chapter 4, page 72 Chapter 4: Syntactic Relations and Case Marking 4.0 General Considerations An important locus of the interaction of syntax, semantics and pragmatics is grammatical relations. RRG takes a rather different view of grammatical relations from other theories. In the first place, it does not consider them to be basic, nor does it derive them from structural configurations. Second, it recognizes only one syntactic function, not three like other theories; there is nothing in RRG corresponding to notions like direct object and indirect object. The syntactic function posited in RRG is not, therefore, part of the same system of oppositions as the traditional notions of grammatical relations (i.e. subject vs. direct object vs. indirect object), and consequently it is not really comparable to the traditional notion that is its closest analog, subject. Third, RRG does not assume that grammatical relations are universal, in two senses. On the one hand, it does not claim that all languages must have grammatical relations in addition to semantic roles, which are universal. On the other hand, in those languages in which a non-semantic grammatical relation can be motivated, the syntactic function posited need not have the same properties in every language; that is, the role of this syntactic function in the grammar of language X may be very different from that played by the syntactic function in language Y, and consequently, the two cannot be considered to be exactly the same. Variation in grammatical relations systems is directly related to differences in the syntax-semantics-pragmatics interface across languages.
    [Show full text]
  • Objects and Obj
    OBJECTS AND OBJ Kersti Börjars and Nigel Vincent The University of Manchester Proceedings of the LFG08 Conference University of Sydney Miriam Butt and Tracy Holloway King (Editors) 2008 CSLI Publications http://csli-publications.stanford.edu/ Abstract The notion of object plays an important role in both descriptive and theoretical work, especially so in a theory such as Lexical-Functional Grammar, where a separate functional structure is assumed in which grammatical relations are captured. In spite of the importance attached to the notion, the object is a relatively understudied phenomenon. In this paper, we consider how the function object links to semantic content, in particular to thematic roles. We conclude that unlike subject, object is not associated with any easily definable semantic content, it is a semantically inert grammatical function. To the extent that it is associated with any one thematic role, this is the Theme, the vaguest of thematic roles. We show how languages exploit this semantic vagueness, for instance through the use of cognate object and pseudo-objects and we consider the impact of this for the association between thematic roles and grammatical relations. 1. Introduction In both typological and theoretical work, the characterization of the core relation object has taken second place to that of subject, with very few studies being devoted exclusively to the properties of objects (Plank (1984b) is an honourable but by now inevitably slightly dated exception).1 Yet, the term ‘object’ has been used in talk about language for many centuries (for a summary see Lepschy 1992). It belongs to a longstanding descriptive and language-teaching tradition which derives its core concepts from the grammar of the classical languages, and especially of Latin.
    [Show full text]
  • A Classifier-Based Approach to Preposition and Determiner Error
    A classifier-based approach to preposition and determiner error correction in L2 English Rachele De Felice and Stephen G. Pulman Oxford University Computing Laboratory Wolfson Building, Parks Road, Oxford OX1 3QD, UK rachele.defelice|stephen.pulman @comlab.ox.ac.uk { } Abstract then present our proposed approach on both L1 and L2 data and discuss the results obtained so far. In this paper, we present an approach to the automatic identification and correction of preposition and determiner errors in non- 2 The problem native (L2) English writing. We show that models of use for these parts of speech 2.1 Prepositions can be learned with an accuracy of 70.06% Prepositions are challenging for learners because and 92.15% respectively on L1 text, and they can appear to have an idiosyncratic behaviour present first results in an error detection which does not follow any predictable pattern even task for L2 writing. across nearly identical contexts. For example, we 1 Introduction say I study in Boston but I study at MIT; or He is independent of his parents, but dependent on his The field of research in natural language process- son. As it is hard even for L1 speakers to articulate ing (NLP) applications for L2 language is con- the reasons for these differences, it is not surpris- stantly growing. This is largely driven by the ex- ing that learners find it difficult to master preposi- panding population of L2 English speakers, whose tions. varying levels of ability may require different types of NLP tools from those designed primarily for native speakers of the language.
    [Show full text]
  • Toward a Balanced Grammatical Description
    Language Documentation & Conservation Special Publication No. 8 (July 2014): The Art and Practice of Grammar Writing, ed. by Toshihide Nakayama and Keren Rice, pp. 91-108 http://nflrc.hawaii.edu/ldc/ 6 http://hdl.handle.net/10125/4586 Toward a balanced grammatical description Thomas E. Payne University of Oregon and SIL International The writer of a grammatical description attempts to accomplish many goals in one complex document. Some of these goals seem to conflict with one another, thus causing tension, discouragement and paralysis for many descriptive linguists. For example, all grammar writers want their work to speak clearly to general linguists and to specialists in their language area tradition. Yet a grammar that addresses universal issues, may not be detailed enough for specialists; while a highly detailed description written in a specialized areal framework may be incomprehensible to those outside of a particular tradition. In the present chapter, I describe four tensions that grammar writers often face, and provide concrete suggestions on how to balance these tensions effectively and creatively. These tensions are: • Comprehensiveness vs. usefulness. • Technical accuracy vs. understandability. • Universality vs. specificity. • A ‘form-driven’ vs. a ‘function-driven’ approach. By drawing attention to these potential conflicts, I hope to help free junior linguists from the unrealistic expectation that their work must fully accomplish all of the ideals that motivate the complex task of describing the grammar of a language. The goal of a description grammar is to produce an esthetically pleasing, intellectually stimulating, and genuinely informative piece of work. 1. INTRODUCTION. Scholars who attempt to write linguistic grammars of underdocu- mented languages strive to accomplish several worthy goals, many of which seem to conflict with one another.1 Tension, discouragement and paralysis may arise as the author of a grammar attempts to fulfill unrealistic expectations of what a grammar ‘should’ be.
    [Show full text]
  • Typological Variation in Grammatical Relations
    Typological variation in grammatical relations Dissertation zur Erlangung des akademischen Grades doctor philosophiae (Dr. phil.) Eingereicht an der Philologischen Fakultät der Universität Leipzig von Alena Witzlack-Makarevich Leipzig, im Juni 2010 ii Contents Acknowledgements .......................... viii Abbreviations ............................. ix 1 Introduction 1 1.1 Grammatical relations ..................... 1 1.2 Goals and overview ....................... 7 2 Major developments in the study of grammatical rela- tions 11 2.1 Introduction ........................... 11 2.2 Anderson (1976) ........................ 11 2.3 Keenan (1976) .......................... 13 2.4 Dixon (1994, 2009) ....................... 14 2.5 Role and Reference Grammar ................ 19 2.6 Radical Construction Grammar ............... 26 2.7 Bickel (in press) and the present approach ........ 28 2.8 Conclusion ............................ 33 3 Methodological background 35 3.1 Introduction ........................... 35 3.2 The goals of modern linguistic typology .......... 35 3.3 Multivariate approach ..................... 36 3.4 Summary and outlook ..................... 39 4 Argument structure and semantic roles 41 4.1 Introduction ........................... 41 4.2 Valence and argument structure ............... 41 4.3 Semantic roles .......................... 48 4.3.1 Role-list approaches to semantic roles ....... 48 4.3.2 Decompositional approaches to semantic roles .. 49 4.3.2.1 Dowty (1991) .................. 50 4.3.2.2 Primus (1999) ................
    [Show full text]
  • The Implementation of Grammatical Functions in Functional Discourse Grammar
    THE IMPLEMENTATION OF GRAMMATICAL FUNCTIONS IN FUNCTIONAL DISCOURSE GRAMMAR Dik BAKKER1 Anna SIEWIERSKA2 • ABSTRACT: In standard FG (DIK, 1997) grammatical functions are assigned directly to the underlying representation in a more or less across the board fashion, only taking into consideration the language dependent semantic function hierarchy. This approach bypasses a number of constraints on subject assignment that may be gathered from typological data, and observed from the actual behaviour of speakers. In this contribution, we make an attempt to reinterpret FG syntactic functions in the light of the FDG model. Following ideas from Givón (1997), we propose a treatment of Subject assignment on the basis of a combination of semantic and pragmatic factors of the relevant referents and other functional aspects of underlying representations. The assignment rules adhere to the respective hierarchies as discussed in the typological literature. In our proposal, Subject (and Object) assignment are now located in the expression component, more specifically in the dynamic version of the expression rules as proposed in Bakker (2001). • KEYWORDS: Subject assignment; alignment; multifactor approach; dynamic expression rules; typological hierarchies. 1 Introduction In the grammar model of Functional Grammar as presented in Dik (1997, p.60) the fully specified underlying clause (FSUC) is an amalgamation of all functional information necessary to derive the morphosyntactic structure of the corresponding expression. The proposition and embedded layers provide the semantics. The illocutionary layer represents the speech act information. Furthermore, all three types of functions, which are crucial for the determination of the shape and order of noun phrases, are coded in the FSUC. Semantic functions are found on all layers.
    [Show full text]
  • Grammatical Relations Typology
    Bickel, B (2010). Grammatical relations typology. In: Song, J J. The Oxford Handbook of Language Typology. Oxford, 399 - 444. ISBN 978-0-19-928125-1. Postprint available at: Zurich Open Repository and Archive http://www.zora.uzh.ch University of Zurich Main Library Posted at the Zurich Open Repository and Archive, University of Zurich. Winterthurerstr. 190 http://www.zora.uzh.ch CH-8057 Zurich www.zora.uzh.ch Originally published at: Bickel, B (2010). Grammatical relations typology. In: Song, J J. The Oxford Handbook of Language Typology. Oxford, 399 - 444. ISBN 978-0-19-928125-1. Year: 2010 Grammatical relations typology Bickel, B Bickel, B (2010). Grammatical relations typology. In: Song, J J. The Oxford Handbook of Language Typology. Oxford, 399 - 444. ISBN 978-0-19-928125-1. Postprint available at: http://www.zora.uzh.ch Posted at the Zurich Open Repository and Archive, University of Zurich. http://www.zora.uzh.ch Originally published at: Bickel, B (2010). Grammatical relations typology. In: Song, J J. The Oxford Handbook of Language Typology. Oxford, 399 - 444. ISBN 978-0-19-928125-1. Draft of a chapter for The Oxford Handbook of Language Typology, ed. Jae Jung Song. Revised version, July 2007 • DRAFT: NOT FOR QUOTATION OR COPYING Grammatical Relations Typology Balthasar Bickel University of Leipzig 1. Grammatical relations past and present Traditionally, the term ‘grammatical relation’ (GR) refers to the morphosyntactic pro- perties that relate an argument to a clause, as, for example, its subject or its object. Alternative terms are ‘syntactic function’ or ‘syntactic role’, and they highlight the fact that GRs are defined by the way in which arguments are integrated syntactically into a clause, i.e.
    [Show full text]
  • Grammatical Relation Extraction in Arabic Language
    Journal of Computer Science 8 (6): 891-898, 2012 ISSN 1549-3636 © 2012 Science Publications Grammatical Relation Extraction in Arabic Language Othman Ibrahim Hammadi and Mohd Juzaiddin Ab Aziz School of Computer Science, Faculty of Information Science and Technology, University Kebangsaan Malaysia, 43600, Selangor, Malaysia Abstract: Problem statement: Grammatical Relation (GR) can be defined as a linguistic relation established by grammar, where linguistic relation is an association among the linguistic forms or constituents. Fundamentally the GR determines grammatical behaviors such as: placement of a word in a clause, verb agreement and the passivity behavior. The GR of Arabic language is a necessary prerequisite for many natural language processing applications, such as machine translation and information retrieval. This study focuses on the GR related problems of Arabic language and addresses the issue with optimum solution. Approach: We had proposed a rule based production method to recognize Grammatical Relations (GRs), as the rule-based approach had been successfully used in developing many natural language processing systems. In order to eradicate the problems of sentence structure recognition, the proposed technique enhances the basic representations of Arabic language such as: Noun Phrase (NP), Verb Phrase (VP), Preposition Phrase (PP) and Adjective Phrase (AP). We had implemented and evaluated the Rule-Based approach that handles chunking and GRs of Arabic sentences. Results: The system was manually tested on 80 Arabic sentences, with the length of each sentence ranging from 3-20 words. The results had yielded the F-score of 83.60%. This outcome proves the viability of this approach for Arabic sentences of GRs extraction.
    [Show full text]
  • Cascaded Grammatical Relation Assignment
    Cascaded Grammatical Relation Assignment Sabine Buchholz and Jorn Veenstra and Walter Daelemans ILK, Computational Linguistics, Tilburg University PO box 90153, 5000 LE Tilburg, The Netherlands [buchholz, veenstra, dae lemans] 0kub. nl Abstract higher modules? In this paper we discuss cascaded Memory- Recently, many people have looked at cas- Based grammatical relations assignment. In the caded and/or shallow parsing and GR assign- first stages of the cascade, we find chunks of sev- ment. Abney (1991) is one of the first who pro- eral types (NP,VP,ADJP,ADVP,PP) and label posed to split up parsing into several cascades. them with their adverbial function (e.g. local, He suggests to first find the chunks and then temporal). In the last stage, we assign gram- the dependecies between these chunks. Grefen- matical relations to pairs of chunks. We stud- stette (1996) describes a cascade of finite-state ied the effect of adding several levels to this cas- transducers, which first finds noun and verb caded classifier and we found that even the less groups, then their heads, and finally syntactic performing chunkers enhanced the performance functions. Brants and Skut (1998) describe a of the relation finder. partially automated annotation tool which con- structs a complete parse of a sentence by recur- 1 Introduction sively adding levels to the tree. (Collins, 1997; When dealing with large amounts of text, find- Ratnaparkhi, 1997) use cascaded processing for ing structure in sentences is often a useful pre- full parsing with good results. Argamon et al. processing step. Traditionally, full parsing is (1998) applied Memory-Based Sequence Learn- used to find structure in sentences.
    [Show full text]