A Resource of Corpus-Derived Typed Predicate Argument Structures for Croatian

Total Page:16

File Type:pdf, Size:1020Kb

A Resource of Corpus-Derived Typed Predicate Argument Structures for Croatian CROATPAS: A Resource of Corpus-derived Typed Predicate Argument Structures for Croatian Costanza Marini Elisabetta Ježek University of Pavia University of Pavia Department of Humanities Department of Humanities costanza.marini01@ [email protected] universitadipavia.it Abstract 2014). The potential uses and purposes of the resource range from multilingual The goal of this paper is to introduce pattern linking between compatible CROATPAS, the Croatian sister project resources to computer-assisted language of the Italian Typed-Predicate Argument learning (CALL). Structure resource (TPAS1, Ježek et al. 2014). CROATPAS is a corpus-based 1 Introduction digital collection of verb valency structures with the addition of semantic Nowadays, we live in a time when digital tools type specifications (SemTypes) to each and resources for language technology are argument slot, which is currently being constantly mushrooming all around the world. developed at the University of Pavia. However, we should remind ourselves that some Salient verbal patterns are discovered languages need our attention more than others if following a lexicographical methodology they are not to face – to put it in Rehm and called Corpus Pattern Analysis (CPA, Hegelesevere’s words – “a steadily increasing Hanks 2004 & 2012; Hanks & and rather severe threat of digital extinction” Pustejovsky 2005; Hanks et al. 2015), (2018: 3282). According to the findings of initiatives such as whereas SemTypes – such as [HUMAN], [ENTITY] or [ANIMAL] – are taken from a the META-NET White Paper Series (Tadić et al. shallow ontology shared by both TPAS 2012; Rehm et al. 2014), we can state that and the Pattern Dictionary of English Croatian is unfortunately among the 21 out of 24 Verbs (PDEV2, Hanks & Pustejovsky official languages of the European Union that are 2005; El Maarouf et al. 2014). The currently considered under-resourced. As a theoretical framework the resource relies matter of fact, Croatian “tools and resources for on is Pustejovsky’s Generative Lexicon […] deep parsing, machine translation, text theory (1995 & 1998; Pustejovsky & semantics, discourse processing, language Ježek 2008), in light of which verbal generation, dialogue management simply do not polysemy and metonymic argument exist” (Tadić et al. 2012: 77). An observation shifts can be traced back to that is only strengthened by the update study compositional operations involving the carried out by Rehm et al. (2014), which shows variation of the SemTypes associated to that, in comparison with other European the valency structure of each verb. The languages, Croatian has weak to no support as corpus used to identify verb patterns in far as text analytics technologies go and only CROATPAS is the Croatian Web as fragmentary support when talking of resources Corpus (hrWac 2.2, RELDI PoS-tagged) such as corpora, lexical resources and grammars. (Ljubešić & Erjavec 2011), which In this framework, a semantic resource such as contains 1.2 billion types and is available CROATPAS could play its part not only in NLP, on the Sketch Engine3 (Kilgarriff et al. (e.g. multilingual pattern linking between other existing compatible resources), but also in automatic machine translation, computer-assisted 1 http://tpas.fbk.eu (last visited on July 12th 2019) 2 http://pdev.org.uk (last visited on July 12th 2019) 3 https://www.sketchengine.eu/ (last visited on July 12th 2019) Copyright © 2019 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). language learning (CALL) and theoretical and our 74 Italian candidates around the median applied cross-linguistic studies. frequency value after taking out the first and the The paper is structured as follows: first a detailed last 20 verbs on the list. Finally, the Croatian overview of the resource is presented (Section 2), translational equivalents for these verbs were followed by its theoretical underpinnings chosen using the 2017 Zanichelli Italian/Croatian (Section 3) and a summary of the Croatian- bilingual dictionary Croato compatto, edited by specific challenges we faced while building the Aleksandra Špikić. resource editor (Section 4). An overview of the The theoretical framework the resource relies on existing related works is given in Section 5. is Pustejovsky’s Generative Lexicon theory Finally, Section 6 hints at the creation of a (1995 & 1998; Pustejovsky & Ježek 2008), in multilingual resource linking CROATPAS, light of which verbal polysemy and metonymic TPAS (Italian) and PDEV (English) patterns and shifts can be traced back to compositional explores CROATPAS’s potential for computer- operations involving the contextual variation of assisted L2 teaching and learning. the SemTypes associated to the valency structure of each verb. 2 Resource overview CROATPAS rests on four key-components, namely: CROATPAS, i.e. the Croatian Typed-Predicate Argument Structure resource, is the Croatian 1) a representative corpus of Croatian; equivalent of the Italian TPAS resource (Ježek et 2) a shallow ontology of SemTypes; al. 2014) and is a corpus-derived collection of 3) a methodology for corpus analysis; Croatian verb argument structures, whose 4) adequate corpus tools. argument slots have been annotated using As for the first component, the corpus used to semantic type specifications (SemTypes). identify verb patterns is the Croatian Web as The first version of the resource is currently Corpus (hrWac 2.2, RELDI PoS-tagged) being developed at the University of Pavia with (Ljubešić & Erjavec, 2011), containing 1.2 the technical assistance of Lexical Computing billion types and available on the Sketch Engine Ltd. in the person of Vìt Baisa and will be (Kilgarriff et al. 2014). We chose to work with released in 2020 through an Open Access the Croatian Web as Corpus since the reference graphical user interface on the website of the corpus for the Italian TPAS resource is a reduced Language Centre of the University of Pavia version of the Italian Web as Corpus (Baroni & (CLA)4. Kilgarriff, 2006), so as to make the two resources CROATPAS contains a sample of 100 medium- as comparable as possible. frequency Croatian verbs, whose Italian As for the shallow ontology of Semantic Type translational counterparts are already available in labels, CROATPAS is based on the same the TPAS resource: 26 of these verbs are hierarchy shared by TPAS and the PDEV project Croatian translational equivalents of Italian of 180 SemTypes, which originates from the “coercive verbs”, i.e. verbs that instantiate Brandeis Shallow Ontology (BSO) (Pustejovsky metonymic shifts in one of their senses (Ježek & et al. 2004) and its initial 65 labels. As pointed Quochi 2010), while the remaining 74 are out by Ježek (2014: 890), SemTypes “are not Croatian translational equivalents of a sample of abstract categories but semantic classes Italian fundamental verbs, i.e. verbs belonging to discovered by generalizing over the statistically that group of approximately 2000 lexemes relevant list of collocates that fill each position”. deemed essential for communicating in Italian For example, the Croatian lexical set for the and that can be found in any sort of text (De SemType [BEVERAGE] in the context of the verb Mauro 2016). pair PITI/POPITI (= TO DRINK, imperfective/perfective) Our 74-verbs sample was selected as follows: we contains, among others: {vodu = water, kavu = first extracted the frequency counts for all the coffee, koktel = cocktail, vino = wine, čaj = tea, 452 fundamental verbs on De Mauro’s list from a pivo = beer, limonadu = lemonade}, as shown in reduced version of the ItWAC (Baroni & the following pattern string from the resource. Kilgarriff, 2006), which contains over 900 million tokens and is available on the Sketch Engine (Kilgarriff et al. 2014). We then selected Figure 1 – One of the pattern strings of PITI 4 https://cla.unipv.it/?page_id=53723 (last visited on July 12th 2019) The corpus analysis methodology used for both and are able to provide different interpretations TPAS and CROATPAS is a lexicographical for a wide range of lexical phenomena. methodology called Corpus Pattern Analysis The principle of co-composition, for instance, (CPA, Hanks 2004 & 2012; Hanks & offers an alternative take on verbal polysemy Pustejovsky 2005; Hanks et al. 2015), which is with respect to traditional accounts. If we based on the Theory of Norms and Exploitations consider lexical items expressing verb arguments (TNE, Hanks 2004, 2013). TNE divides word to be as semantically active and influential as the uses in two main classes: conventional uses verb itself (Pustejovsky 2002: 421), we do not (norms) and deviations from the norms need to think of verbs as polysemous, but rather (exploitations). CPA’s potential lies in that it conceive their meaning as contextually defined does not try to identify meaning in isolation, but by the SemTypes of the surrounding arguments. rather associates it with prototypical contexts, For instance, if we apply this reasoning to the thus focusing on the norms. The standard CPA Croatian verb pair PITI/POPITI (= TO DRINK, procedure requires: imperfective/perfective), we can notice how its meaning changes depending on what is said to be 1) sampling concordances for each verb “drunk”, namely a [BEVERAGE] (1), a [DRUG] 2) identifying its typical patterns – i.e. (2) or a {GOAL} (3). senses – while going through the corpus lines (1) [[HUMANNOM] PIJE [[BEVERAGE]ACC] Djeca ne piju kavu. 3) assigning SemTypes to the argument Children don’t drink
Recommended publications
  • International Computer Science Institute Activity Report 2007
    International Computer Science Institute Activity Report 2007 International Computer Science Institute, 1947 Center Street, Suite 600, Berkeley, CA 94704-1198 USA phone: (510) 666 2900 (510) fax: 666 2956 [email protected] http://www.icsi.berkeley.edu PRINCIPAL 2007 SPONSORS Cisco Defense Advanced Research Projects Agency (DARPA) Disruptive Technology Offi ce (DTO, formerly ARDA) European Union (via University of Edinburgh) Finnish National Technology Agency (TEKES) German Academic Exchange Service (DAAD) Google IM2 National Centre of Competence in Research, Switzerland Microsoft National Science Foundation (NSF) Qualcomm Spanish Ministry of Education and Science (MEC) AFFILIATED 2007 SPONSORS Appscio Ask AT&T Intel National Institutes of Health (NIH) SAP Volkswagen CORPORATE OFFICERS Prof. Nelson Morgan (President and Institute Director) Dr. Marcia Bush (Vice President and Associate Director) Prof. Scott Shenker (Secretary and Treasurer) BOARD OF TRUSTEES, JANUARY 2008 Prof. Javier Aracil, MEC and Universidad Autónoma de Madrid Prof. Hervé Bourlard, IDIAP and EPFL Vice Chancellor Beth Burnside, UC Berkeley Dr. Adele Goldberg, Agile Mind, Inc. and Pharmaceutrix, Inc. Dr. Greg Heinzinger, Qualcomm Mr. Clifford Higgerson, Walden International Prof. Richard Karp, ICSI and UC Berkeley Prof. Nelson Morgan, ICSI (Director) and UC Berkeley Dr. David Nagel, Ascona Group Prof. Prabhakar Raghavan, Stanford and Yahoo! Prof. Stuart Russell, UC Berkeley Computer Science Division Chair Mr. Jouko Salo, TEKES Prof. Shankar Sastry, UC Berkeley, Dean
    [Show full text]
  • 3 Corpus Tools for Lexicographers
    Comp. by: pg0994 Stage : Proof ChapterID: 0001546186 Date:14/5/12 Time:16:20:14 Filepath:d:/womat-filecopy/0001546186.3D31 OUP UNCORRECTED PROOF – FIRST PROOF, 14/5/2012, SPi 3 Corpus tools for lexicographers ADAM KILGARRIFF AND IZTOK KOSEM 3.1 Introduction To analyse corpus data, lexicographers need software that allows them to search, manipulate and save data, a ‘corpus tool’. A good corpus tool is the key to a comprehensive lexicographic analysis—a corpus without a good tool to access it is of little use. Both corpus compilation and corpus tools have been swept along by general technological advances over the last three decades. Compiling and storing corpora has become far faster and easier, so corpora tend to be much larger than previous ones. Most of the first COBUILD dictionary was produced from a corpus of eight million words. Several of the leading English dictionaries of the 1990s were produced using the British National Corpus (BNC), of 100 million words. Current lexico- graphic projects we are involved in use corpora of around a billion words—though this is still less than one hundredth of one percent of the English language text available on the Web (see Rundell, this volume). The amount of data to analyse has thus increased significantly, and corpus tools have had to be improved to assist lexicographers in adapting to this change. Corpus tools have become faster, more multifunctional, and customizable. In the COBUILD project, getting concordance output took a long time and then the concordances were printed on paper and handed out to lexicographers (Clear 1987).
    [Show full text]
  • Better Web Corpora for Corpus Linguistics and NLP
    Masaryk University Faculty of Informatics Better Web Corpora For Corpus Linguistics And NLP Doctoral Thesis Vít Suchomel Brno, Spring 2020 Masaryk University Faculty of Informatics Better Web Corpora For Corpus Linguistics And NLP Doctoral Thesis Vít Suchomel Brno, Spring 2020 Declaration Hereby I declare that this paper is my original authorial work, which I have worked out on my own. All sources, references, and literature used or excerpted during elaboration of this work are properly cited and listed in complete reference to the due source. Vít Suchomel Advisor: Pavel Rychlý i Acknowledgements I would like to thank my advisors, prof. Karel Pala and prof. Pavel Rychlý for their problem insight, help with software design and con- stant encouragement. I am also grateful to my colleagues from Natural Language Process- ing Centre at Masaryk University and Lexical Computing, especially Miloš Jakubíček, Pavel Rychlý and Aleš Horák, for their support of my work and invaluable advice. Furthermore, I would like to thank Adam Kilgarriff, who gave me a wonderful opportunity to work for a leading company in the field of lexicography and corpus driven NLP and Jan Pomikálek who helped me to start. I thank to my wife Kateřina who supported me a lot during writing this thesis. Of those who have always accepted me and loved me in spite of my failures, God is the greatest. ii Abstract The internet is used by computational linguists, lexicographers and social scientists as an immensely large source of text data for various NLP tasks and language studies. Web corpora can be built in sizes which would be virtually impossible to achieve using traditional corpus creation methods.
    [Show full text]
  • Overview: Computational Lexical Semantics and the Week Ahead
    Computational Lexical Semantics My Background and Research Interests Overview: Computational Lexical Semantics and the Week Ahead Diana McCarthy University of Melbourne, July 2011 McCarthy Overview Computational Lexical Semantics My Background and Research Interests Outline 1 Computational Lexical Semantics Word Meaning Representation Distributional Similarity Word Sense Disambiguation Semantic Relations Multiword Expressions Predicate Argument Structure: the syntax-semantics interface 2 My Background and Research Interests Academic Interests Commercial Interests /Demos Sketch Engine, and related tools Dante Other Related Projects McCarthy Overview Word Meaning Representation Computational Lexical Semantics Distributional Similarity My Background and Research Interests Word Sense Disambiguation Semantic Relations Multiword Expressions Motivation Predicate Argument Structure: the syntax-semantics interface we interpret and use language for communication words have meaning if we want to machines to manipulate language as we do they need to be able to distinguish meanings and use words appropriately McCarthy Overview Word Meaning Representation Computational Lexical Semantics Distributional Similarity My Background and Research Interests Word Sense Disambiguation Semantic Relations Multiword Expressions Drawbacks Predicate Argument Structure: the syntax-semantics interface semantic phenomena covert what is the appropriate representation? more variation compared to syntax and morphology less straightforward to evaluate, unless we focus on
    [Show full text]
  • Advances in Computational Linguistics
    Advances in Computational Linguistics Research in Computing Science Series Editorial Board Comité Editorial de la Serie Editors-in-Chief: Associate Editors: Editores en Jefe Editores Asociados Juan Humberto Sossa Azuela (Mexico) Jesús Angulo (France) Gerhard Ritter (USA) Jihad El-Sana (Israel) Jean Serra (France) Jesús Figueroa (Mexico) Ulises Cortés (Spain) Alexander Gelbukh (Russia) Ioannis Kakadiaris (USA) Serguei Levachkine (Russia) Petros Maragos (Greece) Julian Padget (UK) Mateo Valero (Spain) Editorial Coordination: Formatting: Coordinación Editorial Formato Blanca Miranda Valencia Sulema Torres Ramos Research in Computing Science es una publicación trimestral, de circulación internacional, editada por el Centro de Investigación en Computación del IPN, para dar a conocer los avances de investigación científica y desarrollo tecnológico de la comunidad científica internacional. Volumen 41 Febrero, 2009. Tiraje: 500 ejemplares. Certificado de Reserva de Derechos al Uso Exclusivo del Título No. 04-2004-062613250000- 102, expedido por el Instituto Nacional de Derecho de Autor. Certificado de Licitud de Título No. 12897, Certificado de licitud de Contenido No. 10470, expedidos por la Comisión Calificadora de Publicaciones y Revistas Ilustradas. El contenido de los artículos es responsabilidad exclusiva de sus respectivos autores. Queda prohibida la reproducción total o parcial, por cualquier medio, sin el permiso expreso del editor, excepto para uso personal o de estudio haciendo cita explícita en la primera página de cada documento. Imagen de la portada: http://tierramayaimports.com/images/PL3terraA.jpg. Impreso en la Ciudad de México, en los Talleres Gráficos del IPN – Dirección de Publicaciones, Tres Guerras 27, Centro Histórico, México, D.F. Distribuida por el Centro de Investigación en Computación, Av.
    [Show full text]
  • Unit 3: Available Corpora and Software
    Corpus building and investigation for the Humanities: An on-line information pack about corpus investigation techniques for the Humanities Unit 3: Available corpora and software Irina Dahlmann, University of Nottingham 3.1 Commonly-used reference corpora and how to find them This section provides an overview of commonly-used and readily available corpora. It is also intended as a summary only and is far from exhaustive, but should prove useful as a starting point to see what kinds of corpora are available. The Corpora here are divided into the following categories: • Corpora of General English • Monitor corpora • Corpora of Spoken English • Corpora of Academic English • Corpora of Professional English • Corpora of Learner English (First and Second Language Acquisition) • Historical (Diachronic) Corpora of English • Corpora in other languages • Parallel Corpora/Multilingual Corpora Each entry contains the name of the corpus and a hyperlink where further information is available. All the information was accurate at the time of writing but the information is subject to change and further web searches may be required. Corpora of General English The American National Corpus http://www.americannationalcorpus.org/ Size: The first release contains 11.5 million words. The final release will contain 100 million words. Content: Written and Spoken American English. Access/Cost: The second release is available from the Linguistic Data Consortium (http://projects.ldc.upenn.edu/ANC/) for $75. The British National Corpus http://www.natcorp.ox.ac.uk/ Size: 100 million words. Content: Written (90%) and Spoken (10%) British English. Access/Cost: The BNC World Edition is available as both a CD-ROM or via online subscription.
    [Show full text]
  • Adam Kilgarriff's Legacy to Computational Linguistics And
    Adam Kilgarriff’s Legacy to Computational Linguistics and Beyond Roger Evans?, Alexander Gelbukhy, Gregory Grefenstettez, Patrick Hanks}, Miloš Jakubícekˇ ♧, Diana McCarthy⊕, Martha Palmer~, Ted Pedersen/, Michael Rundell], Pavel Rychlý1, Serge Sharoff♤, David Tugwell ? University of Brighton, [email protected], y CIC, Instituto Politécnico Nacional, Mexico, [email protected], z Inria Saclay, [email protected], } University of Wolverhampton, [email protected], ♧ Lexical Computing and Masaryk University, [email protected], ⊕ DTAL University of Cambridge, [email protected], ~ University of Colorado, [email protected], / University of Minnesota, [email protected] ] Lexicography MasterClass, [email protected] 1 Lexical Computing and Masaryk University, pary@fi.muni.cz, ♤ University of Leeds, [email protected], Independent Researcher, [email protected] Abstract. This year, the CICLing conference is dedicated to the memory of Adam Kilgarriff who died last year. Adam leaves behind a tremendous scientific legacy and those working in computational linguistics, other fields of linguistics and lexicography are indebted to him. This paper is a summary review of some of Adam’s main scientific contributions. It is not and cannot be exhaustive. It is writ- ten by only a small selection of his large network of collaborators. Nevertheless we hope this will provide a useful summary for readers wanting to know more about the origins of work, events and software that are so widely relied upon by scientists today, and undoubtedly will continue to be so in the foreseeable future. 1 Introduction Last year was marred by the loss of Adam Kilgarriff who during the last 27 years contributed greatly to the field of computational linguistics1, as well as to other fields of linguistics and to lexicography.
    [Show full text]
  • Finding Definitions in Large Corpora with Sketch Engine
    Finding Definitions in Large Corpora with Sketch Engine Vojtechˇ Kova´rˇ1;2, Monika Mociarikovˇ a´1, Pavel Rychly´1;2 1NLP Centre, Faculty of Informatics, Masaryk University, Brno, Czech Republic 2Lexical Computing CZ, Brno, Czech Republic xkovar3@fi.muni.cz, [email protected], pary@fi.muni.cz Abstract The paper describes automatic definition finding implemented within the leading corpus query and management tool, Sketch Engine. The implementation exploits complex pattern-matching queries in the corpus query language (CQL) and the indexing mechanism of word sketches for finding and storing definition candidates throughout the corpus. The approach is evaluated for Czech and English corpora, showing that the results are usable in practice: precision of the tool ranges between 30 and 75 percent (depending on the major corpus text types) and we were able to extract nearly 2 million definition candidates from an English corpus with 1.4 billion words. The feature is embedded into the interface as a concordance filter, so that users can search for definitions of any query to the corpus, including very specific multi-word queries. The results also indicate that ordinary texts (unlike explanatory texts) contain rather low number of definitions, which is perhaps the most important problem with automatic definition finding in general. Keywords: Sketch Engine, definition, definitions, CQL, corpora 1. Introduction 2. Related Work Google definition boxes1 (that appear sometimes when you Definitions, in the sense of text descriptions of particular search on Google) are very well-known result of “auto- concepts, are an important part of any learning material, matic” definition finding. However, they are probably (to and especially dictionaries.
    [Show full text]
  • A Toolbox for Linguistic Discovery1
    SKETCH ENGINE: A TOOLBOX FOR LINGUISTIC DISCOVERY1 THOMAS, James: DISCOVERING ENGLISH WITH SKETCH ENGINE: A CORPUS- BASED APPROACH TO LANGUAGE EXPLORATION. 2nd ed. Brno: Versatile, 2016. 228 p. ISBN: 9788026083603 THOMAS, James: DISCOVERING ENGLISH WITH SKETCH ENGINE: WORKBOOK AND GLOSSARY. Brno: Versatile, 2016. 140 p. ISBN: 9788026095798 A new paradigm in linguistics, new language analysis tools along with easy access to a substantial amount of electronic texts have made researchers, lexicographers, terminologists, translators and language instructors increasingly interested in corpora. This has led to a higher demand for practical corpus knowledge and skills: which query tools and corpora are available, what they can do and how to use them. However, learning to use query tools requires a greater awareness of the linguistic theory within the neo-Firthian paradigm, which focuses on the interconnection of lexis and grammar. Operational knowledge is inseparable from the theoretical implications that guide research questions and help in the interpretation of corpus data. Indeed, using a corpus tool means understanding and accepting a certain amount of linguistic assumptions incorporated into it by the designers (Anthony, 2013). In language teaching, the gap between scientific advances and the methodologies in use tends to be wide. There is a clear lack of instructional materials and theory-backed know-how in the field. In this context, the new comprehensive textbook onSketch Engine (SkE) by James Thomas is especially timely. This book constitutes a practical hands-on introduction to corpus linguistics built around SkE and focused on English language teaching (ELT) and learning. SkE is a set of software tools for corpus analysis developed by Lexical Computing Ltd.
    [Show full text]
  • Semantic Word Sketches
    The Sketch Engine Semantic Tagging Semantic Tags in Sketch Engine Comparison to FrameNet Conclusions References Semantic Word Sketches Diana McCarthy,† Adam Kilgarriff, MiloˇsJakub´ıˇcek, ‡ Siva Reddy? 3 3 DTAL University of Cambridge†, Lexical Computing , University of Edinburgh?, Masaryk University‡ 3 July 2015 Semantic Word Sketches Outline 1 The Sketch Engine Concordances Word Sketches 2 Semantic Tagging Super Sense Tagger (sst) sst Supersenses 3 Semantic Tags in Sketch Engine In the Concordance Semantic Word Sketches Other Possibilities from sst Output 4 Comparison to FrameNet 5 Conclusions The Sketch Engine Semantic Tagging Semantic Tags in Sketch Engine Concordances Comparison to FrameNet Word Sketches Conclusions References The Sketch Engine concordances, word lists, collocations word sketches create and examine syntactic profiles and collocations of words input automatic part-of-speech tags and a bespoke `sketch grammar' automatic thesauruses: which other words have similar profiles? sketch differences between words Semantic Word Sketches The Sketch Engine Semantic Tagging Semantic Tags in Sketch Engine Concordances Comparison to FrameNet Word Sketches Conclusions References The Sketch Engine for viewing corpora Semantic Word Sketches The Sketch Engine Semantic Tagging Semantic Tags in Sketch Engine Concordances Comparison to FrameNet Word Sketches Conclusions References The Sketch Engine Word Sketches: syntactic profiles Semantic Word Sketches The Sketch Engine Semantic Tagging Semantic Tags in Sketch Engine Concordances Comparison
    [Show full text]
  • Deliverable D3.3.-B Third Batch of Language Resources
    Contract no. 271022 CESAR Central and South-East European Resources Project no. 271022 Deliverable D3.3.-B Third batch of language resources: actions on resources Version No. 1.1 07/02/2013 D3.3-B V 1.1 Page 1 of 81 Contract no. 271022 Document Information Deliverable number: D3.3.-B Deliverable title: Third batch of language resources: actions on resources Due date of deliverable: 31/01/2013 Actual submission date of 08/02/2013 deliverable: Main Author(s): György Szaszák (BME-TMIT) Participants: Mátyás Bartalis, András Balog (BME-TMIT) Željko Agić, Božo Bekavac, Nikola Ljubešić, Sanda Martinčić- Ipšić, Ida Raffaelli, add Jan Šnajder, Krešo Šojat, Vanja Štefanec, Marko Tadić (FFZG) Cvetana Krstev, Duško Vitas, Miloš Utvić, Ranka Stanković, Ivan Obradović, Miljana Milanović, Anđelka Zečević, Mirko Spasić (UBG) Svetla Koeva, Tsvetana Dimitrova, Ivelina Stoyanova (IBL) Tibor Pintér (HASRIL) Mladen Stanojević, Mirko Spasić, Natalija Kovačević, Uroš Milošević, Nikola Dragović, Jelena Jovanović (IPUP) Anna Kibort, Łukasz Kobyliński, Mateusz Kopeć, Katarzyna Krasnowska, Michał Lenart, Marek Maziarz, Marcin Miłkowski, Bartłomiej Nitoń, Agnieszka Patejuk, Aleksander Pohl, Piotr Przybyła, Agata Savary, Filip Skwarski, Jan Szejko, Jakub Waszczuk, Marcin Woliński, Alina Wróblewska, Bartosz Zaborowski, Marcin Zając (IPIPAN) Piotr Pęzik, Łukasz Dróżdż, Maciej Buczek (ULodz) Radovan Garabík, Adriána Žáková (LSIL) Internal reviewer: HASRIL Workpackage: WP3 Workpackage title: Enhancing language resources D3.3-B V 1.1 Page 2 of 81 Contract no. 271022 Workpackage
    [Show full text]
  • John Benjamins Publishing Company
    John Benjamins Publishing Company This is a contribution from Multiple Affordances of Language Corpora for Data-driven Learning. Edited by Agnieszka Leńko-Szymańska and Alex Boulton. © 2015. John Benjamins Publishing Company This electronic file may not be altered in any way. The author(s) of this article is/are permitted to use this PDF file to generate printed copies to be used by way of offprints, for their personal use only. Permission is granted by the publishers to post this file on a closed server which is accessible to members (students and staff) only of the author’s/s’ institute, it is not permitted to post this PDF on the open internet. For any other use of this material prior written permission should be obtained from the publishers or through the Copyright Clearance Center (for USA: www.copyright.com). Please contact [email protected] or consult our website: www.benjamins.com Tables of Contents, abstracts and guidelines are available at www.benjamins.com Stealing a march on collocation Deriving extended collocations from full text for student analysis and synthesis James Thomas Faculty of Arts, Masaryk University Full text affords language learners many opportunities to observe a wide range of linguistic features whose typicality they can ascertain through corpus searches. The particular features investigated in this chapter revolve around the collocations of key words in texts. Given that knowing a collocation in no way guarantees its correct use, a procedure referred to as Collocation Plus has been developed in which learners explore the lexical and grammatical environments of collocations in the contexts in which they meet them.
    [Show full text]