Effective Statistical Models for Syntactic and Semantic Disambiguation

Total Page:16

File Type:pdf, Size:1020Kb

Effective Statistical Models for Syntactic and Semantic Disambiguation EFFECTIVE STATISTICAL MODELS FOR SYNTACTIC AND SEMANTIC DISAMBIGUATION a dissertation submitted to the department of computer science and the committee on graduate studies of stanford university in partial fulfillment of the requirements for the degree of doctor of philosophy Kristina Nikolova Toutanova September 2005 c Copyright by Kristina Nikolova Toutanova 2005 All Rights Reserved ii I certify that I have read this dissertation and that, in my opinion, it is fully adequate in scope and quality as a dissertation for the degree of Doctor of Philosophy. Christopher D. Manning (Principal Adviser) I certify that I have read this dissertation and that, in my opinion, it is fully adequate in scope and quality as a dissertation for the degree of Doctor of Philosophy. Andrew Y. Ng I certify that I have read this dissertation and that, in my opinion, it is fully adequate in scope and quality as a dissertation for the degree of Doctor of Philosophy. Daniel Jurafsky Approved for the University Committee on Graduate Studies. iii iv Abstract This thesis focuses on building effective statistical models for disambiguation of so- phisticated syntactic and semantic natural language (NL) structures. We advance the state of the art in several domains by (i) choosing representations that encode domain knowledge more effectively and (ii) developing machine learning algorithms that deal with the specific properties of NL disambiguation tasks { sparsity of training data and large, structured spaces of hidden labels. For the task of syntactic disambiguation, we propose a novel representation of parse trees that connects the words of the sentence with the hidden syntactic struc- ture in a direct way. Experimental evaluation on parse selection for a Head Driven Phrase Structure Grammar shows the new representation achieves superior perfor- mance compared to previous models. For the task of disambiguating the semantic role structure of verbs, we build a more accurate model, which captures the knowledge that the semantic frame of a verb is a joint structure with strong dependencies be- tween arguments. We achieve this using a Conditional Random Field without Markov independence assumptions on the sequence of semantic role labels. To address the sparsity problem in machine learning for NL, we develop a method for incorporating many additional sources of information, using Markov chains in the space of words. The Markov chain framework makes it possible to combine multiple knowledge sources, to learn how much to trust each of them, and to chain inferences together. It achieves large gains in the task of disambiguating prepositional phrase attachments. v Acknowledgements My thesis work would not have been possible without the help of my advisor, other collaborators, and fellow students. I am especially fortunate to have been advised by Chris Manning. Firstly, I am grateful to him for teaching me almost everything I know about doing research and being part of the academic community. Secondly, I deeply appreciate his constant support and advice on many levels. The work in this thesis was profoundly shaped by numerous insightful discussions with him. I am also very happy to have been able to collaborate with Andrew Ng on random walk models for word dependency distributions. It has been a source of inspiration to interact with someone having such far-reaching research goals and being an endless source of ideas. He also gave me valuable advice on multiple occasions. Many thanks to Dan Jurafsky for initially bringing semantic role labeling to my attention as an interesting domain my research would fit in, contributing useful ideas, and helping with my dissertation on a short notice. Thanks also to the other members of my thesis defense committee { Trevor Hastie and Francine Chen. I am also grateful to Dan Flickinger and Stephan Oepen for numerous discussions on my work in hpsg parsing. Being part of the NLP and the greater AI group at Stanford has been extremely stimulating and fun. I am very happy to have shared an office with Dan Klein and Roger Levy for several years, and with Bill McCartney for one year. Dan taught me, among other things, to always aim to win when entering a competition, and to understand things from first principles. Roger was an example of how to be an excellent researcher while maintaining a balanced life with many outside interests. vi I will miss the heated discussions about research and the fun at NLP lunch. And thanks to Jenny for the Quasi-Newton numerical optimization code. Thanks to Aria Haghighi for collaborating on semantic role labeling models and for staying up very early in the morning to finish writing up papers. I would also like to take the chance to express my gratitude to my undergraduate advisor Galia Angelova and my high-school math teacher Georgi Nikov. Although they did not contribute directly to the current effort, I wouldn't be writing this without them. I am indebted in many ways to Penka Markova { most importantly, for being my friend for many years and for teaching me optimism and self-confidence, and additionally, for collaborating with me on the hpsg parsing work. I will miss the foosball games and green tea breaks with Rajat Raina and Mayur Naik. Thanks also to Jason Townsend and Haiyan Liu for being my good friends. Thanks to Galen Andrew for making my last year at Stanford the happiest. Fi- nally, many thanks to my parents Diana and Nikola, my sister Maria, and my nieces Iana and Diana, for their support, and for bringing me great peace and joy. I gratefully acknowledge the financial support through the ROSIE project funded by Scottish Enterprise under the Stanford-Edinburgh link programme and through ARDA's Advanced Question Answering for Intelligence (AQUAINT) program. vii Contents Abstract v Acknowledgements vi 1 Introduction 1 1.1 A New Representation for Parse Trees . 6 1.1.1 Motivation . 6 1.1.2 Representation . 8 1.1.3 Summary of Results . 8 1.2 Mapping Syntactic to Semantic Structures . 9 1.2.1 General Ideas . 9 1.2.2 Summary of Results . 11 1.3 Lexical Relations . 12 1.3.1 General Ideas . 13 1.3.2 Summary of Results . 14 2 Leaf Path Kernels 16 2.1 Preliminaries . 17 2.1.1 Support Vector Machines . 18 2.1.2 Kernels . 19 2.2 The Redwoods Corpus . 21 viii 2.2.1 Data Characteristics . 24 2.3 Related Work . 25 2.3.1 Penn Treebank Parsing Models . 25 2.3.2 Models for Deep Grammars . 27 2.3.3 Kernels for Parse Trees . 29 2.4 Our Approach . 30 2.5 Proposed Representation . 31 2.5.1 Representing HPSG Signs . 31 2.5.2 The Leaf Projection Paths View of Parse Trees . 32 2.6 Tree and String Kernels . 34 2.6.1 Kernels on Trees Based on Kernels on Projection Paths . 34 2.6.2 String Kernels . 36 2.7 Experiments . 40 2.7.1 SVM Implementation . 41 2.7.2 The Leaf Projection Paths View versus the Context-Free Rule View . 43 2.7.3 Experimental Results using String Kernels on Projection Paths 46 2.7.4 A Note on Application to Penn Treebank Parse Trees . 50 2.8 Discussion and Conclusions . 51 3 Shallow Semantic Parsing 52 3.1 Description of Annotations . 53 3.2 Previous Approaches to Semantic Role Labeling . 56 3.3 Ideas of This Work . 66 3.4 Data and Evaluation Measures . 68 3.5 Local Classifiers . 76 3.5.1 Additional Features for Displaced Constituents . 77 ix 3.5.2 Enforcing the Non-overlapping Constraint . 80 3.5.3 On Split Constituents . 83 3.6 Joint Classifiers . 85 3.6.1 Joint Model Results . 92 3.7 Automatic Parses . 95 3.7.1 Using Multiple Automatic Parse Guesses . 96 3.7.2 Evaluation on the CoNLL 2005 Shared Task . 99 3.8 Conclusions . 102 4 Estimating Word Dependency Distributions 105 4.1 Introduction . 105 4.2 Markov Chain Preliminaries . 108 4.3 Related Work on Smoothing . 111 4.3.1 Estimating Distributions as Limiting Distributions of Markov Chains . 113 4.4 The Prepositional Phrase Attachment Task . 114 4.4.1 Dataset . 115 4.4.2 Previous Work on PP Attachment . 116 4.5 The PP Attachment Model . 119 4.5.1 Baseline Model . 122 4.5.2 Baseline Results . 124 4.6 Random Walks for PP Attachment . 126 4.6.1 Formal Model . 130 4.6.2 Parameter Estimation . 134 4.6.3 Link Types for PP Attachment . 140 4.6.4 Random Walk Experiments . 143 4.6.5 Extended Example . 147 x 4.7 Discussion . 152 4.7.1 Relation to the Transformation Model of Jason Eisner . 152 4.7.2 Relation to Other Models and Conclusions . 154 Bibliography 156 xi List of Figures 1.1 An hpsg derivation and a schematic semantic representation corre- sponding to it for the sentence Stanford offers a distance-learning course in Artificial Intelligence. 2 1.2 A syntactic Penn Treebank-style parse tree and Propbank-style se- mantic annotation for the sentence Stanford offers a distance-learning course in Artificial Intelligence. 3 1.3 An example word space with similarities for the relation in(noun1; noun2). 13 2.1 A grammar rule and lexical entries in a toy unification-based grammar. 22 2.2 Native and derived Redwoods representations for the sentence Do you want to meet on Tuesday? | (a) derivation tree using unique rule and lexical item (in bold) identifiers of the source grammar (top), (b) phrase structure tree labelled with user-defined, parameterizable cat- egory abbreviations (center), and (c) elementary dependency graph extracted from the mrs meaning representation (bottom). 23 2.3 Annotated ambiguous sentences used in experiments: The columns are, from left to right, the total number of sentences, average length, and structural ambiguity.
Recommended publications
  • Lexical Rules in Constraint-Based Grammars
    Lexical Rules in Constraint-based Grammars Ted Briscoe* Ann Copestake t University of Cambridge CSLI, Stanford University Lexical rules have been used to cover a very diverse range of phenomena in constraint-based grammars. Examination of the full range of rules proposed shows that Carpenter's (1991) postu- lated upper bound on the length of list-valued attributes such as SUBCAT in the lexicon cannot be maintained, leading to unrestricted generative capacity in constraint-based formalisms utilizing HPSG-style lexical rules. We argue that it is preferable to subdivide such rules into a class of semiproductive lexically governed genuinely lexical rules, and a class offully productive unary syntactic rules. We develop a restricted approach to lexical rules in a typed default feature structure (TDFS) framework (Lascarides et al. 1995; Lascarides and Copestake 1999), which has enough expressivity to state, for example, rules of verb diathesis alternation, but which does not allow arbitrary manipulation of list-valued features. An interpretation of such lexical rules within a probabilistic version of a TDFS-based linguistic (lexical and grammatical) theory allows us to capture the semiproductive nature of genuinely lexical rules, steering an intermediate course between fully generative or purely abbreviatory rules. We illustrate the utility of this approach with a treatment of dative constructions within a linguistic framework that borrows insights from the constraint-based theories: HPSG, UCG, (Zeevat, Klein, and Calder 1987) and construction grammar (Goldberg 1995). We end by outlin- ing how our approach to lexical rules allows for a treatment of passive and recursive affixation, which are generally assumed to require unrestricted list manipulation operations.
    [Show full text]
  • Tilburg University Evaluation of Lexical Representation Formalisms
    Tilburg University Evaluation of lexical representation formalisms Daelemans, W.M.P.; van der Linden, H.J.B.M. Publication date: 1992 Document Version Publisher's PDF, also known as Version of record Link to publication in Tilburg University Research Portal Citation for published version (APA): Daelemans, W. M. P., & van der Linden, H. J. B. M. (1992). Evaluation of lexical representation formalisms. (ITK Research Memo). Institute for Language Technology and Artifical IntelIigence, Tilburg University. General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. • Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal Take down policy If you believe that this document breaches copyright please contact us providing details, and we will remove access to the work immediately and investigate your claim. Download date: 26. sep. 2021 CBM ~~J R ~`~~ ~J~i 8419 ~ ,~~ ; ~~~ 1992 ~o~~~~~~~~~P~~ IIIInI~uuIIII II 14 J ~ I NINIIIVIII~IINllllll~l 1 INST!TUTE FOR LANGUAGE TECHNOLOGY AND ARTIFICIAL INTELLIGENCE INSTITUUT VOOR TAAL- EN KENNISTECHNOLOGIE ITK Research Memo february 1992 Evaluation of Lexical Representation Formalisms Walter Daelemans 8z Erik-Jan van der Linden No. 14 We wish to thank Hans Bcers with whom we worked on DATR.
    [Show full text]
  • 2018 Canliidocs 10575 2018 Canliidocs 10575 LEGAL KNOWLEDGE and INFORMATION SYSTEMS 2018 Canliidocs 10575
    2018 CanLIIDocs 10575 2018 CanLIIDocs 10575 LEGAL KNOWLEDGE AND INFORMATION SYSTEMS 2018 CanLIIDocs 10575 Frontiers in Artificial Intelligence and Applications The book series Frontiers in Artificial Intelligence and Applications (FAIA) covers all aspects of theoretical and applied Artificial Intelligence research in the form of monographs, selected doctoral dissertations, handbooks and proceedings volumes. The FAIA series contains several sub-series, including ‘Information Modelling and Knowledge Bases’ and ‘Knowledge-Based Intelligent Engineering Systems’. It also includes the biennial European Conference on Artificial Intelligence (ECAI) proceedings volumes, and other EurAI (European Association for Artificial 2018 CanLIIDocs 10575 Intelligence, formerly ECCAI) sponsored publications. The series has become a highly visible platform for the publication and dissemination of original research in this field. Volumes are selected for inclusion by an international editorial board of well-known scholars in the field of AI. All contributions to the volumes in the series have been peer reviewed. The FAIA series is indexed in ACM Digital Library; DBLP; EI Compendex; Google Scholar; Scopus; Web of Science: Conference Proceedings Citation Index – Science (CPCI-S) and Book Citation Index – Science (BKCI-S); Zentralblatt MATH. Series Editors: J. Breuker, N. Guarino, J.N. Kok, J. Liu, R. López de Mántaras, R. Mizoguchi, M. Musen, S.K. Pal and N. Zhong Volume 313 Recently published in this series Vol. 312. T. Endrjukaite, A. Dudko, H. Jaakkola, B. Thalheim, Y. Kiyoki and N. Yoshida (Eds.), Information Modelling and Knowledge Bases XXX Vol. 311. M. Coeckelbergh, J. Loh, M. Funk, J. Seibt and M. Nørskov (Eds.), Envisioning Robots in Society – Power, Politics, and Public Space – Proceedings of Robophilosophy 2018 / TRANSOR 2018 Vol.
    [Show full text]
  • Cv Website.Pdf (151.44
    Aur´elieHerbelot Center for Mind / Brain Sciences, University of Trento [email protected] Research interests Computational linguistics: formal semantics, distributional semantics, reference phenomena, models of quantification and genericity, integration of perception and language in semantics. Education 2006{2010 PhD in computer science, University of Cambridge, UK. Supervisor: Ann Copestake Dissertation Title: Underspecified quantification 2005{2006 MPhil in Computer Speech, Text and Internet Technology, University of Cambridge, UK. Supervisor: Ann Copestake Dissertation Title: Acquiring ontological relationships from corpora 2000{2005 BSc Hons in Information Technology and Computing (1:1), Open University, UK. 1997{1998 MA in Literature and Linguistics (`Mention tr`esbien'), Universit´e de Savoie, France. Awarded with `F´elicitationsdu jury' (praise of the committee). 1994{1997 BA in Literature and Linguistics, Universit´ede Savoie, France. Grants and scholarships 2018 NVIDIA recipient of a hardware grant for a TESLA GPU. 2017 Marie Curie Individual Fellowship (24 months, e158,121.60). Success rate 15%. Project score: 95:6%. 2016 Mozilla `MOSS mission partner' award for my open-source project PeARS (Peer-to-peer Agent for Reciprocated Search). $15,500. 2016 NVIDIA recipient of a hardware grant for a TESLA GPU. 2011-2013 Humboldt Research Fellowship for Postdoctoral Researchers from the Alexander von Humboldt Foundation, Germany. ≈ e65,000. 2005{2010 Full MPhil and PhD scholarships from the Engineering and Physical Sciences Research Council (EPSRC), UK. 1 Positions held / industry work Academic positions 2018- Assistant professor (ricercatore), Centre for Mind/Brain Sciences, University of Trento, Italy. 2017-2018 Marie Curie-Sk lodowska Fellow, Department of Translation and Language Sciences, Universitat Pompeu Fabra, Barcelona, Spain.
    [Show full text]
  • Multiword Expressions a Pain in the Neck For
    This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore. Title Multiword expressions : a pain in the neck for NLP. Sag, Ivan A.; Baldwin, Timothy.; Bond, Francis.; Author(s) Copestake, Ann.; Flickinger, Dan. Sag, I. A., Baldwin, T., Bond, F., Copestake, A., & Flickinger, D. (2002). Multiword expressions: A pain in the neck for NLP. Proceedings of Computational Linguistics Citation and Intelligent Text Processing: Third International Conference: CICLing-2002, Lecture Notes in Computer Science, 2276, 1-15. Date 2002 URL http://hdl.handle.net/10220/6828 © 2002 Springer. This is the author created version of a work that has been peer reviewed and accepted for publication by Proceedings of Computational Linguistics and Intelligent Text Processing: Third International Conference: CICLing-2002, LNCS, Springer. It Rights incorporates referee’s comments but changes resulting from the publishing process, such as copyediting, structural formatting, may not be reflected in this document. The published version is available at: [DOI: http://dx.doi.org/10.1007/3-540-45715-1_1]. Multiword Expressions: A Pain in the Neck for NLP? Ivan A. Sag1, Timothy Baldwin1, Francis Bond2, Ann Copestake3, and Dan Flickinger1 1 CSLI, Ventura Hall, Stanford University Stanford, CA 94305 USA fsag,tbaldwin,[email protected] 2 NTT Communication Science Labs., 2-4 Hikaridai Seika-cho, Soraku-gun, Kyoto, Japan 619-0237 [email protected] 3 University of Cambridge, Computer Laboratory, William Gates Building JJ Thomson Avenue, Cambridge CB3 OFD, UK [email protected] Abstract. Multiword expressions are a key problem for the develop- ment of large-scale, linguistically sound natural language processing tech- nology.
    [Show full text]
  • A Bibliography of Books, Theses, Articles, and Technical Reports in Or on Head-Driven Phrase Structure Grammar
    A Bibliography of Books, Theses, Articles, and Technical Reports in or on Head-Driven Phrase Structure Grammar Compiled by Mike Calcagno, Andreas Kathol, and Carl Pollard Version of September 1993 Adger, David. 1989. Sign-Based Morphology: Applications to Noun-Incorporation. Master's thesis, Uni- versity of Edinburgh. Arnold, Doug and Louise Sadler. 1992. Non-Modifying Adjectives in HPSG. In Working Papers in Language Processing 35. Department of Languages and Linguistics, University of Essex. Backofen, Rolf, Harald Trost, and Hans Uzskoreit. 1991. Linked Typed Feature Formalisms and Termino- logical Knowledge Representation Languages in Natural Language Front Ends. DFKI Research Report 91-28. Balari, Sergio. 1991. Why German is not a Null-Subject Language. CLAUS Report 12, University of the Saarland, Saarbriicken. Balari, Sergio. 1991. Information-Based Linguistics and Head-Driven Phrase Structure. In Miguel Filgueiras, Lufs Damas, Nelma Moreira, and Ana Paula Tomas, eds., Natural Language Processing. Berlin: Springer, 55-101. Balari, Sergio. 1992. Feature Structures, Linguistic Information, and Grammatica/ Theory. Doctoral dissertation, Autonomous University of Barcelona. Balari, Sergio. 1992 .. Agreement and 0-roles: Towards an Account of Null Subjects in HPSG, CLAUS Report 25, University of the Saarland, Saarbriicken. Balari, Sergio. In press. Sujetos nulos en HPSG. In Carlos Martin Vide, ed., Proceedings of the VII Cong~so de Lenguajes N aturales y Lenguajes Forma/es. Barcelona: Vic. Balari, Sergio. In press. Two Types of Agreement. In Catalan Working Papers in Linguistics 2. Au- tonomous University of Barcelona. Bird, Steven. 1990. Constraint-Based Phonology. Doctoral 1issertation, University of Edinburgh. (To be published in Studies in Language, Logic, and Information.) Bird, Steven.
    [Show full text]
  • DATR: a Language for Lexical Knowledge Representation
    DATR: A Language for Lexical Knowledge Representation Roger Evans* Gerald Gazdar t University of Brighton University of Sussex Much recent research on the design of natural language lexicons has made use of nonmonotonic inheritance networks as originally developed for general knowledge representation purposes in Artificial Intelligence. DATR is a simple, spartan language for defining nonmonotonic inher- itance networks with path~value equations, one that has been designed specifically for lexical knowledge representation. In keeping with its intendedly minimalist character, it lacks many of the constructs embodied either in general-purpose knowledge representation languages or in contemporary grammar formalisms. The present paper shows that the language is nonetheless sufficiently expressive to represent concisely the structure of lexical information at a variety of levels of linguistic analysis. The paper provides an informal example-based introduction to DATR and to techniques for its use, including finite-state transduction, the encoding of DA Gs and lexical rules, and the representation of ambiguity and alternation. Sample analyses of phenomena such as inflectional syncretism and verbal subcategorization are given that show how the language can be used to squeeze out redundancy from lexical descriptions. 1. Introduction Irregular lexemes are standardly regular in some respect. Most are just like regular lexemes except that they deviate in one or two characteristics. What is needed is a natural way of saying "this lexeme is regular except for this property." One obvious approach is to use nonmonotonicity and inheritance machinery to capture such lexical irregularity (and subregularity), and much recent research into the design of represen- tation languages for natural language lexicons has thus made use of nonmonotonic inheritance networks (or "semantic nets") as originally developed for more general representation purposes in Artificial Intelligence.
    [Show full text]
  • Chapter 2 the Evolution of HPSG Dan Flickinger Stanford University Carl Pollard Ohio State Universtiy Tom Wasow Stanford University
    Chapter 2 The evolution of HPSG Dan Flickinger Stanford University Carl Pollard Ohio State Universtiy Tom Wasow Stanford University HPSG was developed to express insights from theoretical linguistics in a precise formalism that was computationally tractable. It drew ideas from a wide variety of traditions in linguistics, logic, and computer science. Its chief architects were Carl Pollard and Ivan Sag, and its most direct precursors were Generalized Phrase Structure Grammar and Head Grammar. The theory has been applied in the con- struction of computational systems for the analysis of a variety of languages; a few of these systems have been used in practical applications. This chapter sketches the history of the development and application of the theory. Introduction From its inception in 1983, HPSG was intended to serve as a framework for the formulation and implementation of natural language grammars which are (i) lin- guistically motivated, (ii) formally explicit, and (iii) computationally tractable. These desiderata are reflective of HPSG’s dual origins as an academic linguis- tic theory and as part of an industrial grammar implementation project with an eye toward potential practical applications. Here (i) means that the grammars are intended as scientific theories about the languages in question, and thatthe analyses the grammars give rise to are transparently relatable to the predictions Dan Flickinger, Carl Pollard & Tom Wasow. 2018. The evolution of HPSG. in Stefan Müller, Anne Abeillé, Robert D. Borsley & Jean- Pierre Koenig (eds.), Head-Driven Phrase Structure Grammar: The handbook, 7–40. Berlin: Lan- guage Science Press. DOI:⁇ Dan Flickinger, Carl Pollard & Tom Wasow (empirical consequences) of those theories.
    [Show full text]
  • Applying Natural Language Processing Techniques to Speech Prostheses
    From: AAAI Technical Report FS-96-05. Compilation copyright © 1996, AAAI (www.aaai.org). All rights reserved. Applying Natural Language Processing Techniques to Speech Prostheses Ann Copestake Center for the Study of Language and Information (CSLI) Ventura Hall, Stanford University, Stanford, CA94305-4115 aac@csli,stanford, edu Abstract simply because of the time which is taken but because the delays completely disrupt the usual processes of In this paper, we discuss the application of Natural Language Processing (NLP) techniques to improving turn-taking. Thus the other speaker finds it hard to speechprostheses for people with severe motordisabil- avoid interrupting the prosthesis user. ities. Manypeople whoare unable to speak because of This problem can be alleviated in two ways: by im- physical disability utilize text-to-speech generators as proving the design of the interface (keyboard, head prosthetic devices. However,users of speech prosthe- stick, head pointer, eye tracker etc) or by minimizing ses very often have moregeneral loss of motor control the input that is required for a given output. Wewill and, despite aids such as word prediction, inputting concentrate on the latter aspect here, although there the text is slow and difficult. For typical users, cur- is some interdependence and we will briefly mention rent speechprostheses have output rates whichare less some aspects of this below. than a tenth of the speed of normalspeech. Weare ex- Techniques which have been used for minimizing in- ploring various techniques whichcould improverates, put include the following: without sacrificing flexibility of content. Here we de- scribe the statistical wordprediction techniques used in a communicatordeveloped at CSLIand someexper- Abbreviations, icons and alternative languages iments on improving prediction performance.
    [Show full text]