Cascaded Grammatical Relation Assignment

Total Page:16

File Type:pdf, Size:1020Kb

Cascaded Grammatical Relation Assignment Cascaded Grammatical Relation Assignment Sabine Buchholz and Jorn Veenstra and Walter Daelemans ILK, Computational Linguistics, Tilburg University PO box 90153, 5000 LE Tilburg, The Netherlands [buchholz, veenstra, dae lemans] 0kub. nl Abstract higher modules? In this paper we discuss cascaded Memory- Recently, many people have looked at cas- Based grammatical relations assignment. In the caded and/or shallow parsing and GR assign- first stages of the cascade, we find chunks of sev- ment. Abney (1991) is one of the first who pro- eral types (NP,VP,ADJP,ADVP,PP) and label posed to split up parsing into several cascades. them with their adverbial function (e.g. local, He suggests to first find the chunks and then temporal). In the last stage, we assign gram- the dependecies between these chunks. Grefen- matical relations to pairs of chunks. We stud- stette (1996) describes a cascade of finite-state ied the effect of adding several levels to this cas- transducers, which first finds noun and verb caded classifier and we found that even the less groups, then their heads, and finally syntactic performing chunkers enhanced the performance functions. Brants and Skut (1998) describe a of the relation finder. partially automated annotation tool which con- structs a complete parse of a sentence by recur- 1 Introduction sively adding levels to the tree. (Collins, 1997; When dealing with large amounts of text, find- Ratnaparkhi, 1997) use cascaded processing for ing structure in sentences is often a useful pre- full parsing with good results. Argamon et al. processing step. Traditionally, full parsing is (1998) applied Memory-Based Sequence Learn- used to find structure in sentences. However, ing (MBSL) to NP chunking and subject/object full parsing is a complex task and often pro- identification. However, their subject and ob- vides us with more information then we need. ject finders are independent of their chunker For many tasks detecting only shallow struc- (i.e. not cascaded). tures in a sentence in a fast and reliable way is Drawing from this previous work we will to be preferred over full parsing. For example, explicitly study the effect of adding steps to in information retrieval it can be enough to find the grammatical relations assignment cascade. only simple NPs and VPs in a sentence, for in- Through experiments with cascading several formation extraction we might also want to find classifiers, we will show that even using im- relations between constituents as for example perfect classifiers can improve overall perfor- the subject and object of a verb. mance of the cascaded classifier. We illustrate In this paper we discuss some Memory-Based this claim on the task of finding grammati- (MB) shallow parsing techniques to find labeled cal relations (e.g. subject, object, locative) to chunks and grammatical relations in a sentence. verbs in text. The GR assigner uses several Several MB modules have been developed in sources of information step by step such as sev- previous work, such as: a POS tagger (Daele- eral types of XP chunks (NP, VP, PP, ADJP mans et al., 1996), a chunker (Veenstra, 1998; and ADVP), and adverbial functions assigned Tjong Kim Sang and Veenstra, 1999) and a to these chunks (e.g. temporal, local). Since grammatical relation (GR) assigner (Buchholz, not all of these entities are predicted reliably, it 1998). The questions we will answer in this pa- is the question whether each source leads to an per are: Can we reuse these modules in a cas- improvement of the overall GR assignment. cade of classifiers? What is the effect of cascad- In the rest of this paper we will first briefly de- ing? Will errors at a lower level percolate to scribe Memory-Based Learning in Section 2. In 239 Section 3.1, we discuss the chunking classifiers memory-based approaches to parsing, see (Bod, that we later use as steps in the cascade. Sec- 1992) and (Sekine, 1998). tion 3.2 describes the basic GR classifier. Sec- tion 3.3 presents the architecture and results of 3 Methods and Results the cascaded GR assignment experiments. We In this section we describe the stages of the cas- discuss the results in Section 4 and conclude cade. The very first stage consists of a Memory- with Section 5. Based Part-of-Speech Tagger (MBT) for which we refer to (Daelemans et al., 1996). The 2 Memory-Based Learning next three stages involve determining bound- Memory-Based Learning (MBL) keeps all train- aries and labels of chunks. Chunks are non- ing data in memory and only abstracts at clas- recursive, non-overlapping constituent parts of sification time by extrapolating a class from the sentences (see (Abney, 1991)). First, we si- most similar item(s) in memory. In recent work multaneously chunk sentences into: NP-, VP- Daelemans et al. (1999b) have shown that for , Prep-, ADJP- and APVP-chunks. As these typical natural language processing tasks, this chunks are non-overlapping, no words can be- approach is at an advantage because it also long to more than one chunk, and thus no con- "remembers" exceptional, low-frequency cases flicts can arise. Prep-chunks are the preposi- which are useful to extrapolate from. More- tional part of PPs, thus excluding the nominal over, automatic feature weighting in the similar- part. Then we join a Prep-chunk and one -- ity metric of an MB learner makes the approach or more coordinated -- NP-chunks into a PP- well-suited for domains with large numbers of chunk. Finally, we assign adverbial function features from heterogeneous sources, as it em- (ADVFUNC) labels (e.g. locative or temporal) bodies a smoothing-by-similarity method when to all chunks. data is sparse (Zavrel and Daelemans, 1997). In the last stage of the cascade, we label We have used the following MBL algorithms1: several types of grammatical relations between pairs of words in the sentence. IB1 : A variant of the k-nearest neighbor (k- The data for all our experiments was ex- NN) algorithm. The distance between a tracted from the Penn Treebank II Wall Street test item and each memory item is defined Journal (WSJ) corpus (Marcus et al., 1993). as the number of features for which they For all experiments, we used sections 00-19 as have a different value (overlap metric). training material and 20-24 as test material. IBi-IG : IB1 with information gain (an See Section 4 for results on other train/test set information-theoretic notion measuring the splittings. reduction of uncertainty about the class to For evaluation of our results we use the pre- be predicted when knowing the value of a cision and recall measures. Precision is the per- feature) to weight the cost of a feature value centage of predicted chunks/relations that are mismatch during comparison. actually correct, recall is the percentage of cor- rect chunks/relations that are actually found. IGTree : In.this variant, a decision tree is cre- For convenient comparisons of only one value, ated with features as tests, and ordered ac- we also list the FZ=i value (C.J.van Rijsbergen, cording to the information gain of the fea- tures, as a heuristic approximation of the 1979): (Z2+l)'prec'rec/~2.prec+rec , with/~ = 1 computationally more expensive IB1 vari- 3.1 Chunking ants. In the first experiment described in this section, For more references and information about the task is to segment the sentence into chunks these algorithms we refer to (Daelemans et al., and to assign labels to these chunks. This pro- 1998; Daelemans et al., 1999b). For other cess of chunking and labeling is carried out by assigning a tag to each word in a sentence left- 1For the experiments described in this paper we have to-right. Ramshaw and Marcus (1995) first as- used TiMBL, an MBL software package developed in the ILK-group (Daelemans et al., 1998), TiMBL is available signed a chunk tag to each word in the sentence: from: http://ilk.kub.nl/. I for inside a chunk, O for outside a chunk, and 240 B tbr inside a chunk, but tile preceding word is type precision recall •fl=l in another chunk. As we want to find more than NPchunks 92.5 92.2 92.3 one kind of chunk, we have to [hrther differen- VPchunks 91.9 91.7 91.8 tiate tile IOB tags as to which kind of chunk ADJPchunks 68.4 65.0 66.7 (NP, VP, Prep, ADJP or ADVP) the word is ADVPchunks 78.0 77.9 77.9 ill. With the extended IOB tag set at hand we Prepchunks 95.5 96.7 96.1 can tag the sentence: PPchunks 91.9 92.2 92.0 But/CO [NP the/DT dollar/NN NP] ADVFUNCs 78.0 69.5 73.5 [ADVP later/KB ADVP] [VP rebounded/VBD VP] ,/, [VP finishing/VBG VP] Table h Results of chunking-labeling experi- [ADJP slightly/KB higher/RBK ADJP] ments. NP-,VP-, ADJP-, ADVP- and Prep- [Prep against/IN Prep] [NP the/DT chunks are found sinmltaneously, but for con- yen/NNS NP] [ADJP although/IN ADJP] venience, precision and recall values are given [ADJP slightly/RB lower/J JR ADJP] separately for each type of chunk. [Prep against/IN Prep] [NP the/DT mark/NN NP] ./. experiments, one of these words is always a verb, as: since this yields the most important GRs. The But/CCo the/DTi_NP dollar/NNi_Np other word is the head of the phrase which is later/KBi-ADVP rebounded/VBDi_vP ,/,0 annotated with this grammatical relation in the f inishing/VBGi_vP slightly/KBi-ADVP treebank. A preposition is the head of a PP, higher/RBRi_ADVP against/INi_Prep a noun of an NP and so on.
Recommended publications
  • Chapter 30 HPSG and Lexical Functional Grammar Stephen Wechsler the University of Texas Ash Asudeh University of Rochester & Carleton University
    Chapter 30 HPSG and Lexical Functional Grammar Stephen Wechsler The University of Texas Ash Asudeh University of Rochester & Carleton University This chapter compares two closely related grammatical frameworks, Head-Driven Phrase Structure Grammar (HPSG) and Lexical Functional Grammar (LFG). Among the similarities: both frameworks draw a lexicalist distinction between morphology and syntax, both associate certain words with lexical argument structures, both employ semantic theories based on underspecification, and both are fully explicit and computationally implemented. The two frameworks make available many of the same representational resources. Typical differences between the analyses proffered under the two frameworks can often be traced to concomitant differ- ences of emphasis in the design orientations of their founding formulations: while HPSG’s origins emphasized the formal representation of syntactic locality condi- tions, those of LFG emphasized the formal representation of functional equivalence classes across grammatical structures. Our comparison of the two theories includes a point by point syntactic comparison, after which we turn to an exposition ofGlue Semantics, a theory of semantic composition closely associated with LFG. 1 Introduction Head-Driven Phrase Structure Grammar is similar in many respects to its sister framework, Lexical Functional Grammar or LFG (Bresnan et al. 2016; Dalrymple et al. 2019). Both HPSG and LFG are lexicalist frameworks in the sense that they distinguish between the morphological system that creates words and the syn- tax proper that combines those fully inflected words into phrases and sentences. Stephen Wechsler & Ash Asudeh. 2021. HPSG and Lexical Functional Gram- mar. In Stefan Müller, Anne Abeillé, Robert D. Borsley & Jean- Pierre Koenig (eds.), Head-Driven Phrase Structure Grammar: The handbook.
    [Show full text]
  • Grammatical Relations Typology
    Draft of a chapter for The Oxford Handbook of Language Typology, ed. Jae Jung Song. Grammatical Relations Typology Balthasar Bickel University of Leipzig 1. Grammatical relations past and present Traditionally, the term ‘grammatical relation’ (GR) refers to the morphosyntactic pro- perties that relate an argument to a clause, as, for example, its subject or its object. Alternative terms are ‘syntactic function’ or ‘syntactic role’, and they highlight the fact that GRs are defined by the way in which arguments are integrated syntactically into a clause, i.e. by functioning as subject, object etc. Whatever terminology one prefers, what is crucial about the traditional notion of GRs is (a) that they are identi- fied by syntactic properties, and (b) that they relate an argument to the clause. This differentiates GRs from semantic roles (SRs), also known as thematic roles (θ-roles): SRs are semantic, not syntactic relations, and they hold between arguments and predicates (typically verbs), rather than between arguments and clauses. The difference between GRs and SRs is best visible in such contrasts as Sue killed the shark vs. Sue was killed by the shark: in both cases, the NP Sue is the subject of the clause. But in the active version, the referent of Sue is the agent of ‘kill’, while in the passive version, Sue is the patient of ‘kill.’ The syntactic properties that have traditionally been considered the key identifiers of GRs are the property of triggering verb agreement and the property of beeing assigned a specific case. In our example, Sue triggers third person singular agreement in the verb and this identifies the NP as the subject of the clause.
    [Show full text]
  • From Lexical to Functional Categories: New Foundations for the Study of Language Development
    From Lexical to Functional Categories: New Foundations for the Study of Language Development Cristina Dye (Newcastle University), Yarden Kedar (Beit Berl College) & Barbara Lust (Cornell University) 1. Introduction We will use the term Functional Categories (FCs) in this article in contrast to the term Lexical Categories (LCs) to capture a distinction between the nouns, verbs and adjectives (LCs) exemplified in (1) below versus the underlined determiners (the, an), the verb phrase markers (the auxiliary will and the morphological verb inflection -s), the preposition (with), and the complementizer (that) – all of which constitute FCs in this sentence. (1) The young girl will eat an ice cream that her friend shares with her. All natural languages appear to distinguish FCs and LCs (e.g., Abney, 1987; Chomsky, 1995), and FCs, as a fundamental part of syntax, are thought to be unique to human languages (Hauser, Barner, & O’Donnell, 2007, p. 111). In general, FCs are assumed to carry major grammatical weight in sentence construction, whereas the LCs are assumed to carry major semantic force. However, we must recognize the difficulty in making this distinction. For example, even FCs may carry semantic relevance, and some units appear to be mixed, that is, the small class of pronouns such as her in (1) above which involves limited semantic factors such as gender, and plays a functional role in determining phrase structure. In addition, FCs vary across languages in their range and nature. In spite of the fact that FCs are often viewed as representing a small, closed class word category, it is now recognized that “the number of functional elements in syntax is not easy to estimate..
    [Show full text]
  • Chapter 4: Syntactic Relations and Case Marking
    Chapter 4, page 72 Chapter 4: Syntactic Relations and Case Marking 4.0 General Considerations An important locus of the interaction of syntax, semantics and pragmatics is grammatical relations. RRG takes a rather different view of grammatical relations from other theories. In the first place, it does not consider them to be basic, nor does it derive them from structural configurations. Second, it recognizes only one syntactic function, not three like other theories; there is nothing in RRG corresponding to notions like direct object and indirect object. The syntactic function posited in RRG is not, therefore, part of the same system of oppositions as the traditional notions of grammatical relations (i.e. subject vs. direct object vs. indirect object), and consequently it is not really comparable to the traditional notion that is its closest analog, subject. Third, RRG does not assume that grammatical relations are universal, in two senses. On the one hand, it does not claim that all languages must have grammatical relations in addition to semantic roles, which are universal. On the other hand, in those languages in which a non-semantic grammatical relation can be motivated, the syntactic function posited need not have the same properties in every language; that is, the role of this syntactic function in the grammar of language X may be very different from that played by the syntactic function in language Y, and consequently, the two cannot be considered to be exactly the same. Variation in grammatical relations systems is directly related to differences in the syntax-semantics-pragmatics interface across languages.
    [Show full text]
  • Objects and Obj
    OBJECTS AND OBJ Kersti Börjars and Nigel Vincent The University of Manchester Proceedings of the LFG08 Conference University of Sydney Miriam Butt and Tracy Holloway King (Editors) 2008 CSLI Publications http://csli-publications.stanford.edu/ Abstract The notion of object plays an important role in both descriptive and theoretical work, especially so in a theory such as Lexical-Functional Grammar, where a separate functional structure is assumed in which grammatical relations are captured. In spite of the importance attached to the notion, the object is a relatively understudied phenomenon. In this paper, we consider how the function object links to semantic content, in particular to thematic roles. We conclude that unlike subject, object is not associated with any easily definable semantic content, it is a semantically inert grammatical function. To the extent that it is associated with any one thematic role, this is the Theme, the vaguest of thematic roles. We show how languages exploit this semantic vagueness, for instance through the use of cognate object and pseudo-objects and we consider the impact of this for the association between thematic roles and grammatical relations. 1. Introduction In both typological and theoretical work, the characterization of the core relation object has taken second place to that of subject, with very few studies being devoted exclusively to the properties of objects (Plank (1984b) is an honourable but by now inevitably slightly dated exception).1 Yet, the term ‘object’ has been used in talk about language for many centuries (for a summary see Lepschy 1992). It belongs to a longstanding descriptive and language-teaching tradition which derives its core concepts from the grammar of the classical languages, and especially of Latin.
    [Show full text]
  • A Classifier-Based Approach to Preposition and Determiner Error
    A classifier-based approach to preposition and determiner error correction in L2 English Rachele De Felice and Stephen G. Pulman Oxford University Computing Laboratory Wolfson Building, Parks Road, Oxford OX1 3QD, UK rachele.defelice|stephen.pulman @comlab.ox.ac.uk { } Abstract then present our proposed approach on both L1 and L2 data and discuss the results obtained so far. In this paper, we present an approach to the automatic identification and correction of preposition and determiner errors in non- 2 The problem native (L2) English writing. We show that models of use for these parts of speech 2.1 Prepositions can be learned with an accuracy of 70.06% Prepositions are challenging for learners because and 92.15% respectively on L1 text, and they can appear to have an idiosyncratic behaviour present first results in an error detection which does not follow any predictable pattern even task for L2 writing. across nearly identical contexts. For example, we 1 Introduction say I study in Boston but I study at MIT; or He is independent of his parents, but dependent on his The field of research in natural language process- son. As it is hard even for L1 speakers to articulate ing (NLP) applications for L2 language is con- the reasons for these differences, it is not surpris- stantly growing. This is largely driven by the ex- ing that learners find it difficult to master preposi- panding population of L2 English speakers, whose tions. varying levels of ability may require different types of NLP tools from those designed primarily for native speakers of the language.
    [Show full text]
  • Toward a Balanced Grammatical Description
    Language Documentation & Conservation Special Publication No. 8 (July 2014): The Art and Practice of Grammar Writing, ed. by Toshihide Nakayama and Keren Rice, pp. 91-108 http://nflrc.hawaii.edu/ldc/ 6 http://hdl.handle.net/10125/4586 Toward a balanced grammatical description Thomas E. Payne University of Oregon and SIL International The writer of a grammatical description attempts to accomplish many goals in one complex document. Some of these goals seem to conflict with one another, thus causing tension, discouragement and paralysis for many descriptive linguists. For example, all grammar writers want their work to speak clearly to general linguists and to specialists in their language area tradition. Yet a grammar that addresses universal issues, may not be detailed enough for specialists; while a highly detailed description written in a specialized areal framework may be incomprehensible to those outside of a particular tradition. In the present chapter, I describe four tensions that grammar writers often face, and provide concrete suggestions on how to balance these tensions effectively and creatively. These tensions are: • Comprehensiveness vs. usefulness. • Technical accuracy vs. understandability. • Universality vs. specificity. • A ‘form-driven’ vs. a ‘function-driven’ approach. By drawing attention to these potential conflicts, I hope to help free junior linguists from the unrealistic expectation that their work must fully accomplish all of the ideals that motivate the complex task of describing the grammar of a language. The goal of a description grammar is to produce an esthetically pleasing, intellectually stimulating, and genuinely informative piece of work. 1. INTRODUCTION. Scholars who attempt to write linguistic grammars of underdocu- mented languages strive to accomplish several worthy goals, many of which seem to conflict with one another.1 Tension, discouragement and paralysis may arise as the author of a grammar attempts to fulfill unrealistic expectations of what a grammar ‘should’ be.
    [Show full text]
  • Typological Variation in Grammatical Relations
    Typological variation in grammatical relations Dissertation zur Erlangung des akademischen Grades doctor philosophiae (Dr. phil.) Eingereicht an der Philologischen Fakultät der Universität Leipzig von Alena Witzlack-Makarevich Leipzig, im Juni 2010 ii Contents Acknowledgements .......................... viii Abbreviations ............................. ix 1 Introduction 1 1.1 Grammatical relations ..................... 1 1.2 Goals and overview ....................... 7 2 Major developments in the study of grammatical rela- tions 11 2.1 Introduction ........................... 11 2.2 Anderson (1976) ........................ 11 2.3 Keenan (1976) .......................... 13 2.4 Dixon (1994, 2009) ....................... 14 2.5 Role and Reference Grammar ................ 19 2.6 Radical Construction Grammar ............... 26 2.7 Bickel (in press) and the present approach ........ 28 2.8 Conclusion ............................ 33 3 Methodological background 35 3.1 Introduction ........................... 35 3.2 The goals of modern linguistic typology .......... 35 3.3 Multivariate approach ..................... 36 3.4 Summary and outlook ..................... 39 4 Argument structure and semantic roles 41 4.1 Introduction ........................... 41 4.2 Valence and argument structure ............... 41 4.3 Semantic roles .......................... 48 4.3.1 Role-list approaches to semantic roles ....... 48 4.3.2 Decompositional approaches to semantic roles .. 49 4.3.2.1 Dowty (1991) .................. 50 4.3.2.2 Primus (1999) ................
    [Show full text]
  • The Implementation of Grammatical Functions in Functional Discourse Grammar
    THE IMPLEMENTATION OF GRAMMATICAL FUNCTIONS IN FUNCTIONAL DISCOURSE GRAMMAR Dik BAKKER1 Anna SIEWIERSKA2 • ABSTRACT: In standard FG (DIK, 1997) grammatical functions are assigned directly to the underlying representation in a more or less across the board fashion, only taking into consideration the language dependent semantic function hierarchy. This approach bypasses a number of constraints on subject assignment that may be gathered from typological data, and observed from the actual behaviour of speakers. In this contribution, we make an attempt to reinterpret FG syntactic functions in the light of the FDG model. Following ideas from Givón (1997), we propose a treatment of Subject assignment on the basis of a combination of semantic and pragmatic factors of the relevant referents and other functional aspects of underlying representations. The assignment rules adhere to the respective hierarchies as discussed in the typological literature. In our proposal, Subject (and Object) assignment are now located in the expression component, more specifically in the dynamic version of the expression rules as proposed in Bakker (2001). • KEYWORDS: Subject assignment; alignment; multifactor approach; dynamic expression rules; typological hierarchies. 1 Introduction In the grammar model of Functional Grammar as presented in Dik (1997, p.60) the fully specified underlying clause (FSUC) is an amalgamation of all functional information necessary to derive the morphosyntactic structure of the corresponding expression. The proposition and embedded layers provide the semantics. The illocutionary layer represents the speech act information. Furthermore, all three types of functions, which are crucial for the determination of the shape and order of noun phrases, are coded in the FSUC. Semantic functions are found on all layers.
    [Show full text]
  • Grammatical Relations Typology
    Bickel, B (2010). Grammatical relations typology. In: Song, J J. The Oxford Handbook of Language Typology. Oxford, 399 - 444. ISBN 978-0-19-928125-1. Postprint available at: Zurich Open Repository and Archive http://www.zora.uzh.ch University of Zurich Main Library Posted at the Zurich Open Repository and Archive, University of Zurich. Winterthurerstr. 190 http://www.zora.uzh.ch CH-8057 Zurich www.zora.uzh.ch Originally published at: Bickel, B (2010). Grammatical relations typology. In: Song, J J. The Oxford Handbook of Language Typology. Oxford, 399 - 444. ISBN 978-0-19-928125-1. Year: 2010 Grammatical relations typology Bickel, B Bickel, B (2010). Grammatical relations typology. In: Song, J J. The Oxford Handbook of Language Typology. Oxford, 399 - 444. ISBN 978-0-19-928125-1. Postprint available at: http://www.zora.uzh.ch Posted at the Zurich Open Repository and Archive, University of Zurich. http://www.zora.uzh.ch Originally published at: Bickel, B (2010). Grammatical relations typology. In: Song, J J. The Oxford Handbook of Language Typology. Oxford, 399 - 444. ISBN 978-0-19-928125-1. Draft of a chapter for The Oxford Handbook of Language Typology, ed. Jae Jung Song. Revised version, July 2007 • DRAFT: NOT FOR QUOTATION OR COPYING Grammatical Relations Typology Balthasar Bickel University of Leipzig 1. Grammatical relations past and present Traditionally, the term ‘grammatical relation’ (GR) refers to the morphosyntactic pro- perties that relate an argument to a clause, as, for example, its subject or its object. Alternative terms are ‘syntactic function’ or ‘syntactic role’, and they highlight the fact that GRs are defined by the way in which arguments are integrated syntactically into a clause, i.e.
    [Show full text]
  • Grammatical Relation Extraction in Arabic Language
    Journal of Computer Science 8 (6): 891-898, 2012 ISSN 1549-3636 © 2012 Science Publications Grammatical Relation Extraction in Arabic Language Othman Ibrahim Hammadi and Mohd Juzaiddin Ab Aziz School of Computer Science, Faculty of Information Science and Technology, University Kebangsaan Malaysia, 43600, Selangor, Malaysia Abstract: Problem statement: Grammatical Relation (GR) can be defined as a linguistic relation established by grammar, where linguistic relation is an association among the linguistic forms or constituents. Fundamentally the GR determines grammatical behaviors such as: placement of a word in a clause, verb agreement and the passivity behavior. The GR of Arabic language is a necessary prerequisite for many natural language processing applications, such as machine translation and information retrieval. This study focuses on the GR related problems of Arabic language and addresses the issue with optimum solution. Approach: We had proposed a rule based production method to recognize Grammatical Relations (GRs), as the rule-based approach had been successfully used in developing many natural language processing systems. In order to eradicate the problems of sentence structure recognition, the proposed technique enhances the basic representations of Arabic language such as: Noun Phrase (NP), Verb Phrase (VP), Preposition Phrase (PP) and Adjective Phrase (AP). We had implemented and evaluated the Rule-Based approach that handles chunking and GRs of Arabic sentences. Results: The system was manually tested on 80 Arabic sentences, with the length of each sentence ranging from 3-20 words. The results had yielded the F-score of 83.60%. This outcome proves the viability of this approach for Arabic sentences of GRs extraction.
    [Show full text]
  • The Acquisition of Verbal Inflection in Child Grammars in a Variability Model of Early Morphosyntactic Development : a Biolinguistic Perspective
    THE ACQUISITION OF VERBAL INFLECTION IN CHILD GRAMMARS IN A VARIABILITY MODEL OF EARLY MORPHOSYNTACTIC DEVELOPMENT : A BIOLINGUISTIC PERSPECTIVE A Dissertation submitted to the Faculty of the Graduate School of Arts and Sciences of Georgetown University in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Linguistics By Dominik Rus, B.A. Washington, DC April 11, 2008 Copyright 2010 by Dominik Rus All Rights Reserved ii THE ACQUISITION OF VERBAL INFLECTION IN CHILD GRAMMARS IN A VARIABILITY MODEL OF EARLY MORPHOSYNTACTIC DEVELOPMENT : A BIOLINGUISTIC PERSPECTIVE Dominik Rus, B.A. Thesis Advisors: Donna Lardiere, Ph.D. and Charles Yang, Ph.D. ABSTRACT This dissertation investigates the acquisition of early verb inflection in child Slovenian from morphosyntactic and morphophonological perspectives. It centers on the phenomenon of root nonfinites, particularly the patterns of omission and substitution errors in verb inflection marking. It argues that every acquisition model needs to account for the following robust developmental phenomena: initial telegraphic speech, optionality, variability, graded (rather than absolute) morpheme order, and gradualness. It suggests that current biolinguistic models can be enhanced by supplementing evidence from UG-based studies with that from usage-based and neuropsychological accounts. This proposal is tied to Chomsky’s (2005) hypothesis that language comprises three factors —the innate language faculty, experience, and computational efficiency, and that language acquisition relies considerably on this third factor. The study motivates the Inflectional Hierarchy Complexity Hypothesis (IHCH), suggesting that while children’s early morphosyntax is adultlike regarding the availability of iii functional categories and the concatenation operation, children’s morphophonological spell-out is unreliable due to computational bottlenecks.
    [Show full text]