Learning Noun-Modifier Semantic Relations with Corpus-Based

Total Page:16

File Type:pdf, Size:1020Kb

Learning Noun-Modifier Semantic Relations with Corpus-Based Learning Noun-Modifier Semantic Relations with Corpus-based and WordNet-based Features Vivi Nastase Marina Sokolova∗ Stan Szpakowicz Jelber Sayyad-Shirabad D´epartement d’informatique et de School of Information Technology and Engineering School of Information Technology and Engineering recherche op´erationnelle University of Ottawa 800 King Edward Avenue University of Ottawa 800 King Edward Avenue Universit´ede Montr´eal Ottawa, Ontario K1N 6N5, Canada Ottawa, Ontario K1N 6N5, Canada 2920, chemin de la Tour Institute of Computer Science vnastase,jsayyad @site.uottawa.ca Montr´eal, Qu´ebec H3T 1J4, Canada Polish Academy of Sciences { } [email protected] Ordona 21, 01-237 Warszawa, Poland [email protected] Abstract We experiment with two methods of representing the words in a base NP, to be used in Machine Learning (ML) We study the performance of two representations of word experiments for learning semantic relations between nouns meaning in learning noun-modifier semantic relations. One and their modifiers. One method is based on features ex- representation is based on lexical resources, in particular WordNet, the other – on a corpus. We experimented with de- tracted from WordNet, which was designed to capture, de- cision trees, instance-based learning and Support Vector Ma- scribe, and relate word senses. In brief, we use hypernyms chines. All these methods work well in this learning task. We to describe in more and more general terms the sense of report high precision, recall and F-score, and small variation a word in a pair. The other method is based on contex- in performance across several 10-fold cross-validation runs. tual information extracted from corpora. Contexts are use- The corpus-based method has the advantage of working with ful for determining word senses, as reflected in research on data without word-sense annotations and performs well over word-sense disambiguation using a window of words sur- the baseline. The WordNet-based method, requiring word- rounding the target word (starting with (Yarowsky 1995), sense annotated data, has higher precision. and more recently (Purandare & Pedersen 2004)). We use grammatical collocations3 (as opposed to proximity-based Introduction co-occurrences) extracted from the British National Corpus, to describe each word in a pair. We compare the learning In understanding a text, it is essential to recognize rela- results produced by using these two types of word sense de- tions among occurrences1, entities and their attributes, rep- scriptions with the results obtained by (Turney & Littman resented at the surface as verbs, nouns and their modifiers. 2003) and (Turney 2005), who used a paraphrase method to Semantic relations describe interactions between a noun and describe the pair as a whole. its modifiers (noun-modifier relations), a verb and its argu- The data we work with consist of noun-modifier pairs la- ments (case relations/semantic roles), and two clauses. In belled with 30 fine-grained semantic relations, grouped into the past few years we have seen the Natural Language Pro- five relation classes. Experiments presented in this paper are cessing (NLP) community’s renewed interest in analyzing based on the five-class coarse-grained grouping. semantic relations especially between verbs and their argu- The corpus-based method gives precision, recall and F- ments (Baker, Fillmore, & Lowe 1998), (Kipper, Dang, & scores well above the baseline (discussed in the “Experi- Palmer 2000), nouns and their modifiers (Rosario & Hearst mental setup and results” section), and it works on data with- 2001). Semantic role labelling competitions2 also seem to out word-sense annotations. The method based on WordNet increase the attractiveness of this topic. gives results with higher precision, but requires word-sense In this paper we consider a specific supervised learning annotated data. task: assign semantic relations to noun-modifier pairs in The paper continues with related work on learning noun- base noun phrases (base NPs), composed only of a noun modifier relations, description of the data and the represen- and its modifier. To identify such noun-modifier relations tations used, presentation of the experimental setup and the we can rely only on semantic and morphological informa- results, and concludes with a discussion. tion about words themselves. For example, in the base NPs iron gate, brick house, plastic container: iron, brick, plas- Related Work tic are substances, and gate, house, container are artifacts; this suggests a MATERIAL relation in these pairs. On the We focus on methods that analyze and learn semantic rela- other hand, analyzing case relation or clause-level relations tions in noun-phrases. is assisted by prepositions, subordinators, coordinators and Levi (1978) analyzes the formation of nominal com- maybe more elaborate syntactic structures. pounds. She identifies 9 recoverable deletable predicates (RDPs): be, cause, have, make, use, about, for, from, in, Copyright c 2006, American Association for Artificial Intelli- which, when erased from a more complex expression, gen- gence (www.aaai.org). All rights reserved. erate a noun phrase. Levi writes that relations expressed by ∗This work was done at the University of Ottawa. the RDPs may be universal, because from a semantic point 1Occurrence encompasses all types of events, actions, activi- ties, processes, states and accomplishments (Allen 1984) 3Grammatical collocations are collocates that appear with the 2CONLL 2004, 2005 – Semantic Role Labelling shared task target word in a grammatical relation, such as subject, object, http://www.lsi.upc.edu/ srlconll/ . prepositional complement (Kilgarriff et al. 2004). 781 of view they appear to be quite primitive. Different seman- semantic relation, using a 2-nearest-neighbour algorithm. tic relations may be associated with each RDP. For example: This has been generalized in Turney (2005) using what cause – causes/is caused by; have – possession/possessor; he calls Latent Relational Analysis (LRA). For each word make – physically producing/material. in a dataset of pairs, Lin’s thesaurus (Lin 1998) gives a set Berland & Charniak (1999) and Hearst (1992) work with of possible synonyms. All original pairs and pairs generated specific relations, part of and type of respectively. The Au- from synonyms are used to mine a corpus for paraphrases. tomatic Content Extraction project is a research program in All paraphrases are gathered and a few thousand of the most information extraction that focuses on detecting specific re- frequent ones are selected. The selected paraphrases, the lations (such as employer-organization, agent-artifact) be- original word pairs and the synonym pairs are used to build tween seven types of entities (such as person, organization, an incidence matrix, whose dimensionality is reduced using facility) in texts (Zhao & Grishman 2005). Several types singular value decomposition (Landauer & Dumais 1997). of information – lexical, grammatical and contextual – are Similarity between pairs combines scores for similarity be- combined using kernel methods. tween the original word pair and pairs built using synonyms. Vanderwende (1994) uses hand-crafted rules and a dictio- Because we use the same data as Turney & Littman nary built from texts to find clues about the semantic rela- (2003) and Turney (2005), we compare the results of tions in which a word may be involved. This was tested with learning noun-modifier relations using WordNet-based and 97 pairs extracted from the Brown corpus, with an accuracy corpus-based representations, with the results obtained us- of 52%. ing paraphrase-based information. Several systems use lexical resources (domain-specific like MeSH or general like WordNet or Roget’s Thesaurus) Data and Representations to find the appropriate level of generalization for words in a pair, so that words linked by different relations are properly Lists of semantic relations in use range from general, as separated. the lists in PropBank (Palmer, Gildea, & Kingsbury 2005) and NomBank (Myers et al. 2004), to more and more spe- Rosario & Hearst (2001) learn noun-modifiersemantic re- cific, as in VerbNet (Kipper, Dang, & Palmer 2000) and lations in a medical domain, using neural networks. The list FrameNet (Baker, Fillmore, & Lowe 1998), to domain- of 13 relations is tailored to the application domain. Rosario, specific (Rosario & Hearst 2001). The data we use consist Hearst, & Fillmore (2002), continuing that research, look of 600 noun-modifier pairs, tagged with 30 semantic rela- manually for rules which classify correctly noun compounds tions, grouped into 5 classes of relations by general similar- in the medical domain, based on the MeSH lexical hierar- ity (Barker 1998), (Nastase & Szpakowicz 2003), (Turney & chy. The data are extracted automatically from biomedical Littman 2003): journal articles, and sampled for manual analysis. MeSH is traversed top-down to find a level at which the noun com- 1. CAUSAL groups relations enabling or opposing an occur- pounds in different relations are properly separated. rence. Examples (H denotes the head of a base NP, M Lauer (1995) maps words in noun compounds onto cat- denotes the modifier): egories in Roget’s Thesaurus, in order to find probabilities cause - H causes M: flu virus; of occurrence of certain noun compounds and their para- phrases. There is no automatic process to find the best level effect - H is the effect (was caused by) M: exam anxiety; of generalization. Nastase & Szpakowicz (2003) use the purpose - H is
Recommended publications
  • International Computer Science Institute Activity Report 2007
    International Computer Science Institute Activity Report 2007 International Computer Science Institute, 1947 Center Street, Suite 600, Berkeley, CA 94704-1198 USA phone: (510) 666 2900 (510) fax: 666 2956 [email protected] http://www.icsi.berkeley.edu PRINCIPAL 2007 SPONSORS Cisco Defense Advanced Research Projects Agency (DARPA) Disruptive Technology Offi ce (DTO, formerly ARDA) European Union (via University of Edinburgh) Finnish National Technology Agency (TEKES) German Academic Exchange Service (DAAD) Google IM2 National Centre of Competence in Research, Switzerland Microsoft National Science Foundation (NSF) Qualcomm Spanish Ministry of Education and Science (MEC) AFFILIATED 2007 SPONSORS Appscio Ask AT&T Intel National Institutes of Health (NIH) SAP Volkswagen CORPORATE OFFICERS Prof. Nelson Morgan (President and Institute Director) Dr. Marcia Bush (Vice President and Associate Director) Prof. Scott Shenker (Secretary and Treasurer) BOARD OF TRUSTEES, JANUARY 2008 Prof. Javier Aracil, MEC and Universidad Autónoma de Madrid Prof. Hervé Bourlard, IDIAP and EPFL Vice Chancellor Beth Burnside, UC Berkeley Dr. Adele Goldberg, Agile Mind, Inc. and Pharmaceutrix, Inc. Dr. Greg Heinzinger, Qualcomm Mr. Clifford Higgerson, Walden International Prof. Richard Karp, ICSI and UC Berkeley Prof. Nelson Morgan, ICSI (Director) and UC Berkeley Dr. David Nagel, Ascona Group Prof. Prabhakar Raghavan, Stanford and Yahoo! Prof. Stuart Russell, UC Berkeley Computer Science Division Chair Mr. Jouko Salo, TEKES Prof. Shankar Sastry, UC Berkeley, Dean
    [Show full text]
  • A Resource of Corpus-Derived Typed Predicate Argument Structures for Croatian
    CROATPAS: A Resource of Corpus-derived Typed Predicate Argument Structures for Croatian Costanza Marini Elisabetta Ježek University of Pavia University of Pavia Department of Humanities Department of Humanities costanza.marini01@ [email protected] universitadipavia.it Abstract 2014). The potential uses and purposes of the resource range from multilingual The goal of this paper is to introduce pattern linking between compatible CROATPAS, the Croatian sister project resources to computer-assisted language of the Italian Typed-Predicate Argument learning (CALL). Structure resource (TPAS1, Ježek et al. 2014). CROATPAS is a corpus-based 1 Introduction digital collection of verb valency structures with the addition of semantic Nowadays, we live in a time when digital tools type specifications (SemTypes) to each and resources for language technology are argument slot, which is currently being constantly mushrooming all around the world. developed at the University of Pavia. However, we should remind ourselves that some Salient verbal patterns are discovered languages need our attention more than others if following a lexicographical methodology they are not to face – to put it in Rehm and called Corpus Pattern Analysis (CPA, Hegelesevere’s words – “a steadily increasing Hanks 2004 & 2012; Hanks & and rather severe threat of digital extinction” Pustejovsky 2005; Hanks et al. 2015), (2018: 3282). According to the findings of initiatives such as whereas SemTypes – such as [HUMAN], [ENTITY] or [ANIMAL] – are taken from a the META-NET White Paper Series (Tadić et al. shallow ontology shared by both TPAS 2012; Rehm et al. 2014), we can state that and the Pattern Dictionary of English Croatian is unfortunately among the 21 out of 24 Verbs (PDEV2, Hanks & Pustejovsky official languages of the European Union that are 2005; El Maarouf et al.
    [Show full text]
  • 3 Corpus Tools for Lexicographers
    Comp. by: pg0994 Stage : Proof ChapterID: 0001546186 Date:14/5/12 Time:16:20:14 Filepath:d:/womat-filecopy/0001546186.3D31 OUP UNCORRECTED PROOF – FIRST PROOF, 14/5/2012, SPi 3 Corpus tools for lexicographers ADAM KILGARRIFF AND IZTOK KOSEM 3.1 Introduction To analyse corpus data, lexicographers need software that allows them to search, manipulate and save data, a ‘corpus tool’. A good corpus tool is the key to a comprehensive lexicographic analysis—a corpus without a good tool to access it is of little use. Both corpus compilation and corpus tools have been swept along by general technological advances over the last three decades. Compiling and storing corpora has become far faster and easier, so corpora tend to be much larger than previous ones. Most of the first COBUILD dictionary was produced from a corpus of eight million words. Several of the leading English dictionaries of the 1990s were produced using the British National Corpus (BNC), of 100 million words. Current lexico- graphic projects we are involved in use corpora of around a billion words—though this is still less than one hundredth of one percent of the English language text available on the Web (see Rundell, this volume). The amount of data to analyse has thus increased significantly, and corpus tools have had to be improved to assist lexicographers in adapting to this change. Corpus tools have become faster, more multifunctional, and customizable. In the COBUILD project, getting concordance output took a long time and then the concordances were printed on paper and handed out to lexicographers (Clear 1987).
    [Show full text]
  • Better Web Corpora for Corpus Linguistics and NLP
    Masaryk University Faculty of Informatics Better Web Corpora For Corpus Linguistics And NLP Doctoral Thesis Vít Suchomel Brno, Spring 2020 Masaryk University Faculty of Informatics Better Web Corpora For Corpus Linguistics And NLP Doctoral Thesis Vít Suchomel Brno, Spring 2020 Declaration Hereby I declare that this paper is my original authorial work, which I have worked out on my own. All sources, references, and literature used or excerpted during elaboration of this work are properly cited and listed in complete reference to the due source. Vít Suchomel Advisor: Pavel Rychlý i Acknowledgements I would like to thank my advisors, prof. Karel Pala and prof. Pavel Rychlý for their problem insight, help with software design and con- stant encouragement. I am also grateful to my colleagues from Natural Language Process- ing Centre at Masaryk University and Lexical Computing, especially Miloš Jakubíček, Pavel Rychlý and Aleš Horák, for their support of my work and invaluable advice. Furthermore, I would like to thank Adam Kilgarriff, who gave me a wonderful opportunity to work for a leading company in the field of lexicography and corpus driven NLP and Jan Pomikálek who helped me to start. I thank to my wife Kateřina who supported me a lot during writing this thesis. Of those who have always accepted me and loved me in spite of my failures, God is the greatest. ii Abstract The internet is used by computational linguists, lexicographers and social scientists as an immensely large source of text data for various NLP tasks and language studies. Web corpora can be built in sizes which would be virtually impossible to achieve using traditional corpus creation methods.
    [Show full text]
  • Overview: Computational Lexical Semantics and the Week Ahead
    Computational Lexical Semantics My Background and Research Interests Overview: Computational Lexical Semantics and the Week Ahead Diana McCarthy University of Melbourne, July 2011 McCarthy Overview Computational Lexical Semantics My Background and Research Interests Outline 1 Computational Lexical Semantics Word Meaning Representation Distributional Similarity Word Sense Disambiguation Semantic Relations Multiword Expressions Predicate Argument Structure: the syntax-semantics interface 2 My Background and Research Interests Academic Interests Commercial Interests /Demos Sketch Engine, and related tools Dante Other Related Projects McCarthy Overview Word Meaning Representation Computational Lexical Semantics Distributional Similarity My Background and Research Interests Word Sense Disambiguation Semantic Relations Multiword Expressions Motivation Predicate Argument Structure: the syntax-semantics interface we interpret and use language for communication words have meaning if we want to machines to manipulate language as we do they need to be able to distinguish meanings and use words appropriately McCarthy Overview Word Meaning Representation Computational Lexical Semantics Distributional Similarity My Background and Research Interests Word Sense Disambiguation Semantic Relations Multiword Expressions Drawbacks Predicate Argument Structure: the syntax-semantics interface semantic phenomena covert what is the appropriate representation? more variation compared to syntax and morphology less straightforward to evaluate, unless we focus on
    [Show full text]
  • Advances in Computational Linguistics
    Advances in Computational Linguistics Research in Computing Science Series Editorial Board Comité Editorial de la Serie Editors-in-Chief: Associate Editors: Editores en Jefe Editores Asociados Juan Humberto Sossa Azuela (Mexico) Jesús Angulo (France) Gerhard Ritter (USA) Jihad El-Sana (Israel) Jean Serra (France) Jesús Figueroa (Mexico) Ulises Cortés (Spain) Alexander Gelbukh (Russia) Ioannis Kakadiaris (USA) Serguei Levachkine (Russia) Petros Maragos (Greece) Julian Padget (UK) Mateo Valero (Spain) Editorial Coordination: Formatting: Coordinación Editorial Formato Blanca Miranda Valencia Sulema Torres Ramos Research in Computing Science es una publicación trimestral, de circulación internacional, editada por el Centro de Investigación en Computación del IPN, para dar a conocer los avances de investigación científica y desarrollo tecnológico de la comunidad científica internacional. Volumen 41 Febrero, 2009. Tiraje: 500 ejemplares. Certificado de Reserva de Derechos al Uso Exclusivo del Título No. 04-2004-062613250000- 102, expedido por el Instituto Nacional de Derecho de Autor. Certificado de Licitud de Título No. 12897, Certificado de licitud de Contenido No. 10470, expedidos por la Comisión Calificadora de Publicaciones y Revistas Ilustradas. El contenido de los artículos es responsabilidad exclusiva de sus respectivos autores. Queda prohibida la reproducción total o parcial, por cualquier medio, sin el permiso expreso del editor, excepto para uso personal o de estudio haciendo cita explícita en la primera página de cada documento. Imagen de la portada: http://tierramayaimports.com/images/PL3terraA.jpg. Impreso en la Ciudad de México, en los Talleres Gráficos del IPN – Dirección de Publicaciones, Tres Guerras 27, Centro Histórico, México, D.F. Distribuida por el Centro de Investigación en Computación, Av.
    [Show full text]
  • Unit 3: Available Corpora and Software
    Corpus building and investigation for the Humanities: An on-line information pack about corpus investigation techniques for the Humanities Unit 3: Available corpora and software Irina Dahlmann, University of Nottingham 3.1 Commonly-used reference corpora and how to find them This section provides an overview of commonly-used and readily available corpora. It is also intended as a summary only and is far from exhaustive, but should prove useful as a starting point to see what kinds of corpora are available. The Corpora here are divided into the following categories: • Corpora of General English • Monitor corpora • Corpora of Spoken English • Corpora of Academic English • Corpora of Professional English • Corpora of Learner English (First and Second Language Acquisition) • Historical (Diachronic) Corpora of English • Corpora in other languages • Parallel Corpora/Multilingual Corpora Each entry contains the name of the corpus and a hyperlink where further information is available. All the information was accurate at the time of writing but the information is subject to change and further web searches may be required. Corpora of General English The American National Corpus http://www.americannationalcorpus.org/ Size: The first release contains 11.5 million words. The final release will contain 100 million words. Content: Written and Spoken American English. Access/Cost: The second release is available from the Linguistic Data Consortium (http://projects.ldc.upenn.edu/ANC/) for $75. The British National Corpus http://www.natcorp.ox.ac.uk/ Size: 100 million words. Content: Written (90%) and Spoken (10%) British English. Access/Cost: The BNC World Edition is available as both a CD-ROM or via online subscription.
    [Show full text]
  • Adam Kilgarriff's Legacy to Computational Linguistics And
    Adam Kilgarriff’s Legacy to Computational Linguistics and Beyond Roger Evans?, Alexander Gelbukhy, Gregory Grefenstettez, Patrick Hanks}, Miloš Jakubícekˇ ♧, Diana McCarthy⊕, Martha Palmer~, Ted Pedersen/, Michael Rundell], Pavel Rychlý1, Serge Sharoff♤, David Tugwell ? University of Brighton, [email protected], y CIC, Instituto Politécnico Nacional, Mexico, [email protected], z Inria Saclay, [email protected], } University of Wolverhampton, [email protected], ♧ Lexical Computing and Masaryk University, [email protected], ⊕ DTAL University of Cambridge, [email protected], ~ University of Colorado, [email protected], / University of Minnesota, [email protected] ] Lexicography MasterClass, [email protected] 1 Lexical Computing and Masaryk University, pary@fi.muni.cz, ♤ University of Leeds, [email protected], Independent Researcher, [email protected] Abstract. This year, the CICLing conference is dedicated to the memory of Adam Kilgarriff who died last year. Adam leaves behind a tremendous scientific legacy and those working in computational linguistics, other fields of linguistics and lexicography are indebted to him. This paper is a summary review of some of Adam’s main scientific contributions. It is not and cannot be exhaustive. It is writ- ten by only a small selection of his large network of collaborators. Nevertheless we hope this will provide a useful summary for readers wanting to know more about the origins of work, events and software that are so widely relied upon by scientists today, and undoubtedly will continue to be so in the foreseeable future. 1 Introduction Last year was marred by the loss of Adam Kilgarriff who during the last 27 years contributed greatly to the field of computational linguistics1, as well as to other fields of linguistics and to lexicography.
    [Show full text]
  • Finding Definitions in Large Corpora with Sketch Engine
    Finding Definitions in Large Corpora with Sketch Engine Vojtechˇ Kova´rˇ1;2, Monika Mociarikovˇ a´1, Pavel Rychly´1;2 1NLP Centre, Faculty of Informatics, Masaryk University, Brno, Czech Republic 2Lexical Computing CZ, Brno, Czech Republic xkovar3@fi.muni.cz, [email protected], pary@fi.muni.cz Abstract The paper describes automatic definition finding implemented within the leading corpus query and management tool, Sketch Engine. The implementation exploits complex pattern-matching queries in the corpus query language (CQL) and the indexing mechanism of word sketches for finding and storing definition candidates throughout the corpus. The approach is evaluated for Czech and English corpora, showing that the results are usable in practice: precision of the tool ranges between 30 and 75 percent (depending on the major corpus text types) and we were able to extract nearly 2 million definition candidates from an English corpus with 1.4 billion words. The feature is embedded into the interface as a concordance filter, so that users can search for definitions of any query to the corpus, including very specific multi-word queries. The results also indicate that ordinary texts (unlike explanatory texts) contain rather low number of definitions, which is perhaps the most important problem with automatic definition finding in general. Keywords: Sketch Engine, definition, definitions, CQL, corpora 1. Introduction 2. Related Work Google definition boxes1 (that appear sometimes when you Definitions, in the sense of text descriptions of particular search on Google) are very well-known result of “auto- concepts, are an important part of any learning material, matic” definition finding. However, they are probably (to and especially dictionaries.
    [Show full text]
  • A Toolbox for Linguistic Discovery1
    SKETCH ENGINE: A TOOLBOX FOR LINGUISTIC DISCOVERY1 THOMAS, James: DISCOVERING ENGLISH WITH SKETCH ENGINE: A CORPUS- BASED APPROACH TO LANGUAGE EXPLORATION. 2nd ed. Brno: Versatile, 2016. 228 p. ISBN: 9788026083603 THOMAS, James: DISCOVERING ENGLISH WITH SKETCH ENGINE: WORKBOOK AND GLOSSARY. Brno: Versatile, 2016. 140 p. ISBN: 9788026095798 A new paradigm in linguistics, new language analysis tools along with easy access to a substantial amount of electronic texts have made researchers, lexicographers, terminologists, translators and language instructors increasingly interested in corpora. This has led to a higher demand for practical corpus knowledge and skills: which query tools and corpora are available, what they can do and how to use them. However, learning to use query tools requires a greater awareness of the linguistic theory within the neo-Firthian paradigm, which focuses on the interconnection of lexis and grammar. Operational knowledge is inseparable from the theoretical implications that guide research questions and help in the interpretation of corpus data. Indeed, using a corpus tool means understanding and accepting a certain amount of linguistic assumptions incorporated into it by the designers (Anthony, 2013). In language teaching, the gap between scientific advances and the methodologies in use tends to be wide. There is a clear lack of instructional materials and theory-backed know-how in the field. In this context, the new comprehensive textbook onSketch Engine (SkE) by James Thomas is especially timely. This book constitutes a practical hands-on introduction to corpus linguistics built around SkE and focused on English language teaching (ELT) and learning. SkE is a set of software tools for corpus analysis developed by Lexical Computing Ltd.
    [Show full text]
  • Semantic Word Sketches
    The Sketch Engine Semantic Tagging Semantic Tags in Sketch Engine Comparison to FrameNet Conclusions References Semantic Word Sketches Diana McCarthy,† Adam Kilgarriff, MiloˇsJakub´ıˇcek, ‡ Siva Reddy? 3 3 DTAL University of Cambridge†, Lexical Computing , University of Edinburgh?, Masaryk University‡ 3 July 2015 Semantic Word Sketches Outline 1 The Sketch Engine Concordances Word Sketches 2 Semantic Tagging Super Sense Tagger (sst) sst Supersenses 3 Semantic Tags in Sketch Engine In the Concordance Semantic Word Sketches Other Possibilities from sst Output 4 Comparison to FrameNet 5 Conclusions The Sketch Engine Semantic Tagging Semantic Tags in Sketch Engine Concordances Comparison to FrameNet Word Sketches Conclusions References The Sketch Engine concordances, word lists, collocations word sketches create and examine syntactic profiles and collocations of words input automatic part-of-speech tags and a bespoke `sketch grammar' automatic thesauruses: which other words have similar profiles? sketch differences between words Semantic Word Sketches The Sketch Engine Semantic Tagging Semantic Tags in Sketch Engine Concordances Comparison to FrameNet Word Sketches Conclusions References The Sketch Engine for viewing corpora Semantic Word Sketches The Sketch Engine Semantic Tagging Semantic Tags in Sketch Engine Concordances Comparison to FrameNet Word Sketches Conclusions References The Sketch Engine Word Sketches: syntactic profiles Semantic Word Sketches The Sketch Engine Semantic Tagging Semantic Tags in Sketch Engine Concordances Comparison
    [Show full text]
  • Deliverable D3.3.-B Third Batch of Language Resources
    Contract no. 271022 CESAR Central and South-East European Resources Project no. 271022 Deliverable D3.3.-B Third batch of language resources: actions on resources Version No. 1.1 07/02/2013 D3.3-B V 1.1 Page 1 of 81 Contract no. 271022 Document Information Deliverable number: D3.3.-B Deliverable title: Third batch of language resources: actions on resources Due date of deliverable: 31/01/2013 Actual submission date of 08/02/2013 deliverable: Main Author(s): György Szaszák (BME-TMIT) Participants: Mátyás Bartalis, András Balog (BME-TMIT) Željko Agić, Božo Bekavac, Nikola Ljubešić, Sanda Martinčić- Ipšić, Ida Raffaelli, add Jan Šnajder, Krešo Šojat, Vanja Štefanec, Marko Tadić (FFZG) Cvetana Krstev, Duško Vitas, Miloš Utvić, Ranka Stanković, Ivan Obradović, Miljana Milanović, Anđelka Zečević, Mirko Spasić (UBG) Svetla Koeva, Tsvetana Dimitrova, Ivelina Stoyanova (IBL) Tibor Pintér (HASRIL) Mladen Stanojević, Mirko Spasić, Natalija Kovačević, Uroš Milošević, Nikola Dragović, Jelena Jovanović (IPUP) Anna Kibort, Łukasz Kobyliński, Mateusz Kopeć, Katarzyna Krasnowska, Michał Lenart, Marek Maziarz, Marcin Miłkowski, Bartłomiej Nitoń, Agnieszka Patejuk, Aleksander Pohl, Piotr Przybyła, Agata Savary, Filip Skwarski, Jan Szejko, Jakub Waszczuk, Marcin Woliński, Alina Wróblewska, Bartosz Zaborowski, Marcin Zając (IPIPAN) Piotr Pęzik, Łukasz Dróżdż, Maciej Buczek (ULodz) Radovan Garabík, Adriána Žáková (LSIL) Internal reviewer: HASRIL Workpackage: WP3 Workpackage title: Enhancing language resources D3.3-B V 1.1 Page 2 of 81 Contract no. 271022 Workpackage
    [Show full text]