On Statistical Parameter Setting

On Statistical Parameter Setting

On Statistical Parameter Setting Damir ĆAVAR, Joshua HERRING, Giancarlo SCHREMENTI Toshikazu IKUTA, Paul RODRIGUES Computer Science, Indiana University Linguistics Dept., Indiana University Bloomington, IN, 47405 Bloomington, IN, 46405 [email protected] [email protected] properties of a small subset of languages can be Abstract induced with high accuracy, most of the existing We present a model and an experimental approaches are motivated by applied or platform of a bootstrapping approach to engineering concerns, and thus make assumptions statistical induction of natural language that are less cognitively plausible: a. Large corpora properties that is constraint based with voting are processed all at once, though unsupervised components. The system is incremental and incremental induction of grammars is rather the unsupervised. In the following discussion we approach that would be relevant from a focus on the components for morphological psycholinguistic perspective; b. Arbitrary decisions induction. We show that the much harder about selections of sets of elements are made, 2 problem of incremental unsupervised based on frequency or frequency profile rank, morphological induction can outperform though such decisions should rather be derived or comparable all-at-once algorithms with avoided in general. respect to precision. We discuss how we use However, the most important aspects missing in such systems to identify cues for induction in these approaches, however, are the link to different a cross-level architecture. linguistic levels and the support of a general learning model that makes predictions about how 1 Introduction knowledge is induced on different linguistic levels and what the dependencies between information at In recent years there has been a growing amount these levels are. Further, there is no study focusing of work focusing on the computational modeling on the type of supervision that might be necessary of language processing and acquisition, implying a for the guidance of different algorithm types cognitive and theoretical relevance both of the towards grammars that resemble theoretical and models as such, as well as of the language empirical facts about language acquisition, and properties extracted from raw linguistic data.1 In processing and the final knowledge of language. the computational linguistic literature several While many theoretical models of language attempts to induce grammar or linguistic acquisition use innateness as a crutch to avoid knowledge from such data have shown that at outstanding difficulties, both on the general and different levels a high amount of information can abstract level of I-language as well as the more be extracted, even with no or minimal supervision. detailed level of E-language, (see, among others, Different approaches tried to show how various Lightfoot (1999) and Fodor and Teller (2000), puzzles of language induction could be solved. there is also significant research being done which From this perspective, language acquisition is the shows that children take advantage of statistical process of segmentation of non-discrete acoustic regularities in the input for use in the language- input, mapping of segments to symbolic learning task (see Batchelder (1997) and related representations, mapping representations on references within). higher-level representations such as phonology, In language acquisition theories the dominant morphology and syntax, and even induction of view is that knowledge of one linguistic level is semantic properties. Due to space restrictions, we bootstrapped from knowledge of one, or even cannot discuss all these approaches in detail. We several different levels. Just to mention such will focus on the close domain of morphology. approaches: Grimshaw (1981), and Pinker (1984) Approaches to the induction of morphology as presented in e.g. Schone and Jurafsky (2001) or 2 Just to mention some of the arbitrary decisions Goldsmith (2001) show that the morphological made in various approaches, e.g. Mintz (1996) selects a small set of all words, the most frequent words, to induce word types via clustering ; Schone and Jurafsky 1 See Batchelder (1998) for a discussion of these (2001) select words with frequency higher than 5 to aspects. induce morphological segmentation. 9 assume that semantic properties are used to are the cues that are used for grammar induction bootstrap syntactic knowledge, and Mazuka (1998) only. suggested that prosodic properties of language As shown in Elghamry (2004) and Elghamry and establish a bias for specific syntactic properties, Ćavar (2004), there are efficient ways to identify a e.g. headedness or branching direction of kernel set of such units in an unsupervised fashion constituents. However, these approaches are based without any arbitrary decision where to cut the set on conceptual considerations and psycholinguistc of elements and on the basis of what kind of empirical grounds, the formal models and features. They present an algorithm that selects the computational experiments are missing. It is set of kernel cues on the lexical and syntactic level, unclear how the induction processes across as the smallest set of words that co-occurs with all linguistic domains might work algorithmically, and other words. Using this set of words it is possible the quantitative experiments on large scale data are to cluster the lexical inventory into open and missing. closed class words, as well as to identify the As for algorithmic approaches to cross-level subclasses of nouns and verbs in the open class. induction, the best example of an initial attempt to The direction of the selectional preferences of the exploit cues from one level to induce properties of language is derived as an average of point-wise another is presented in Déjean (1998), where Mutual Information on each side of the identified morphological cues are identified for induction of cues and types, which is a self-supervision aspect syntactic structure. Along these lines, we will that biases the search direction for a specific argue for a model of statistical cue-based learning, language. This resulting information is understood introducing a view on bootstrapping as proposed in as derivation of secondary cues, which then can be Elghamry (2004), and Elghamry and Ćavar (2004), used to induce selectional properties of verbs that relies on identification of elementary cues in (frames), as shown in Elghamry (2004). the language input and incremental induction and The general claim thus is: further cue identification across all linguistic • Cues can be identified in an unsupervised levels. fashion in the input. 1.1 Cue-based learning • These cues can be used to induce properties of the target grammar. Presupposing input driven learning, it has been • These properties represent cues that can be shown in the literature that initial segmenations used to induce further cues, and so on. into words (or word-like units) is possible with The hypothesis is that this snowball effect can unsupervised methods (e.g. Brent and Cartwright reduce the search space of the target grammar (1996)), that induction of morphology is possible incrementally. The main research questions are (e.g. Goldsmith (2001), Schone and Jurafsky now, to what extend do different algorithms (2001)) and even the induction of syntactic provide cues for other linguistic levels and what structures (e.g. Van Zaanen (2001)). As mentioned kind of information do they require as supervision earlier, the main drawback of these approaches is in the system, in order to gain the highest accuracy the lack of incrementality, certain arbitrary at each linguistic level, and how does the linguistic decisions about the properties of elements taken information of one level contribute to the into account, and the lack of integration into a information on another. general model of bootstrapping across linguistic In the following, the architectural considerations levels. of such a computational model are discussed, As proposed in Elghamry (2004), cues are resulting in an example implementation that is elementary language units that can be identified at applied to morphology induction, where each linguistic level, dependent or independent of morphological properties are understood to prior induction processes. That is, intrinsic represent cues for lexical clustering as well as properties of elements like segments, syllables, syntactic structure, and vice versa, similar to the morphemes, words, phrases etc. are the ones ideas formulated in Déjean (1998), among others. available for induction procedures. Intrinsic properties are for example the frequency of these 1.2 Incremental Induction Architecture units, their size, and the number of other units they The basic architectural principle we presuppose are build of. Extrinsic properties are taken into is incrementality, where incrementally utterances account as well, where extrinsic stands for are processed. The basic language unit is an distributional properties, the context, relations to utterance, with clear prosodic breaks before and other units of the same type on one, as well as after. The induction algorithm consumes such across linguistic levels. In this model, extrinsic and utterances and breaks them into basic linguistic intrinsic properties of elementary language units units, generating for each step hypotheses about 10 the linguistic structure of each utterance, based on representations for the input in a first order the grammar built so far and statistical properties

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    8 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us