2 the Corpus As Object: Design and Purpose

Total Page:16

File Type:pdf, Size:1020Kb

2 the Corpus As Object: Design and Purpose 2 The corpus as object: Design and purpose As corpora have become larger and more diverse, and as they are more frequently used to make de®nitive statements about language, issues of how they are designed have become more important. Four aspects of corpus design are discussed in this chapter: size, content, representativeness and permanence. The chapter also summarises some types of corpus investigation, each of which treats the corpus as a different kind of object. Issues in corpus design Size As computer technology has advanced since the 1960s, so it has become feasible to store and access corpora of ever-increasing size. Whereas the LOB Corpus and Brown Corpus seemed as big as anyone would ever want, at the time, nowadays 1 million words is fairly small in terms of corpora. The British National Corpus is 100 million words; the Bank of English is currently about 400 million. CANCODE is 5 million words. The feasible size of a corpus is not limited so much by the capacity of a computer to store it, as by the speed and ef®ciency of the access software. If, for example, counting the number of past and present tense forms of the verb BE in a given corpus takes longer than a few minutes, the researcher may prefer to use a smaller corpus whose results might be considered to be just as reliable but on which the software would work much more speedily. Sheer quantity of information can be overwhelming for the observer. In the Bank of English corpus, for example, a search for the word point gives almost 143,000 hits. Few researchers can obtain useful information simply by looking at so many concordance lines. One solution is to use a smaller corpus, which will give less data. Another is to keep the large corpus, but to use software which will make a selection of data from the whole, either by selecting at random a proportion of the total concordance lines, or by identifying and allowing selection of the most signi®cant collocates, or other 25 26 Corpora in applied linguistics signi®cant features. In this way, the data from a very large corpus can be reduced to a manageable scale whilst retaining the advantages of coverage of the large corpus. The question of corpus size can be a contentious one. There are many arguments to support the view that a small corpus can be valuable under certain circumstances. Carter and McCarthy 41995: 143), for example, argue that for the purposes of studying grammar in spoken language a relatively small corpus is suf®cient 4because grammar words tend to be very frequent). Some kinds of corpus investigation that depend on the manual annotation of the corpus are also, necessarily, restricted in terms of the size of the corpus that can feasibly be annotated. Similarly, many commercially available con- cordance software packages restrict the size of the corpus they can be used with 4see Biber et al 1998: 285±286 for details). Someone aiming to build a balanced corpus may restrict the amount of data from one source in order to match data from another source 4for example, the amount of written data may be kept smaller than it need be because it is dif®cult to obtain large amounts of spoken data). A more extreme view is that for some purposes there is an optimum corpus size that should not be exceeded even if practical considerations would allow it, that is, that a relatively small corpus is not just suf®cient but also necessary. However, this argument usually refers to the dif®culties of processing large amounts of information about very frequent words, as discussed above 4e.g. Carter and McCarthy 1995: 143). The opposing view would be that it is preferable to select from a large amount of data than to restrict the amount of data available 4Sinclair 1992). Arguments about optimum corpus size tend to be academic for most people. Most corpus users simply make use of as much data as is available, without worrying too much about what is not available. As well as the very large, general corpora designed to assist in writing dictionaries and other reference books, there are thousands of smaller corpora around the world, some comprising only a few thousand words and designed for a particular piece of research. Content It is a truism that a corpus is neither good nor bad in itself, but suited or not suited to a particular purpose. Decisions about what should go into a corpus are based on what the corpus is going to be used for, but also about what is available. A researcher who wishes to study The corpus as object: Design and purpose 27 academic articles in History, for example, may plan to build a corpus consisting simply of published articles. Even with such a straight- forward design plan, however, a number of issues will have to be resolved, such as: . Will the articles come from one journal, or from a range of journals? How will the journal4s) be selected? . Will the articles chosen be restricted to those apparently written by native speakers of the target language, or not? . Will articles other than research articles, such as book reviews, be included? . If the articles are to be stored electronically, and copyright clearance is required, which publishers are willing to give permis- sion for their articles to be used? . If the articles are required in electronic form, which publishers are willing to make electronic versions of their journals available? To the ®rst three of these questions, there are no `right' answers. The best choice in each case will depend on the precise purpose of the research. The last two questions refer to purely pragmatic issues, but these may in the end affect the design of the corpus more profoundly than the ®rst three. Publishers' policies are not the only issues affecting availability for corpora. Parallel corpora, for example, are composed of texts that are translations of each other. They can make use only of texts that have been translated between the two languages involved. Other issues of selection come second to the question of what translations exist. Using unpublished material in a corpus does not obviate the issue of availability. For example, a researcher wishing to study student writing may ask students to submit electronic versions of their essays to assist in this work. If this submission is voluntary, the researcher will collect only the essays whose authors are willing for their essays to be used. For some purposes, corpus design is scarcely an issue. For example, if a corpus is being used to encourage learners to investigate language data for themselves, the precise contents of that corpus may be relatively unimportant. Any collection of newspapers written in the target form of the language, for example, will be adequate for some investigation purposes, even if the newspapers do not represent all aspects of language use 4see, for example, Dodd 1997). Where the corpus is being taken as representative of a language or a language variety, the notion of design and balance is very important. This will be considered separately below. 28 Corpora in applied linguistics Balance and representativeness Corpora are very often intended to be representative of a particular kind of language. If the object of study is academic prose, or casual conversation, or the language of newspapers, or American English, for example, an attempt must be made to build a corpus that is representative of the whole. 4See Kennedy 1991: 52; Crowdy 1993; Burnard ed. 1995 for discussions of representativeness in the British National Corpus.) Usually, this involves breaking the whole down into component parts and aiming to include equal amounts of data from each of the parts. For example, under `the language of news- papers', it may be decided to include a range of newspaper types 4broadsheet and tabloid, for instance) and a range of article types 4hard news, human interest, editorials, letters, sports, business, advertisements and so on). A balanced corpus might be said to consist of equal numbers of words in each category: broadsheet hard news; tabloid hard news; broadsheet human interest; tabloid human interest and so on. This notion of balance is, however, open to question. It might be argued that broadsheet newspapers contain more words, on average, than tabloids, and that therefore the corpus should contain more from the broadsheets than from the tabloids. Alternatively, it might be argued that tabloids are read by more people than broadsheets and that therefore they should comprise a larger part of the corpus. If the broadsheets contain more hard news and editorial than the tabloids, and the tabloids contain more human interest and sports news, should the proportions in the corpus be adjusted accordingly? In the case of newspapers, the solution may be to include all issues of a selection of publications from a given week, month or year. This will allow the proportions to determine them- selves. It does mean, however, that more data will probably be gathered from one type of publication than another. Similar problems with relation to the collection of academic prose or casual conver- sation, for example, are not open to so easy a solution. The problem is that `being representative' inevitably involves knowing what the character of the `whole' is. Where the proportions of that character are unknowable, attempts to be representative tend to rest on little more than guesswork. The problem of representation and balance becomes more dif®cult when the corpus is supposed to represent a regional variety of English 4Singapore English, or British English, or Australian English), with all its complexity of internal variation.
Recommended publications
  • Corpus Studies in Applied Linguistics
    106 Pietilä, P. & O-P. Salo (eds.) 1999. Multiple Languages – Multiple Perspectives. AFinLA Yearbook 1999. Publications de l’Association Finlandaise de Linguistique Appliquée 57. pp. 105–134. CORPUS STUDIES IN APPLIED LINGUISTICS Kay Wikberg, University of Oslo Stig Johansson, University of Oslo Anna-Brita Stenström, University of Bergen Tuija Virtanen, University of Växjö Three samples of corpora and corpus-based research of great interest to applied linguistics are presented in this paper. The first is the Bergen Corpus of London Teenage Language, a project which has already resulted in a number of investigations of how young Londoners use their language. It has also given rise to a related Nordic project, UNO, and to the project EVA, which aims at developing material for assessing the English proficiency of pupils in the compulsory school system in Norway. The second corpus is the English-Norwegian Parallel Corpus (Oslo), which has provided data for both contrastive studies and the study of translationese. Altogether it consists of about 2.6 million words and now also includes translations of English texts into German, Dutch and Portuguese. The third corpus, the International Corpus of Learner English, is a collection of advanced EFL essays written by learners representing 15 different mother tongues. By comparing linguistic features in the various subcorpora it is possible to find out about non-nativeness generally and about problems shared by students representing different languages. 1 INTRODUCTION 1.1 Corpus studies and descriptive linguistics Corpus-based language research has now been with us for more than 20 years. The number of recent books dealing with corpus studies (cf.
    [Show full text]
  • Download (237Kb)
    CHAPTER II REVIEW OF RELATED LITERATURE In this chapter, the researcher presents the result of reviewing related literature which covers Corpus based analysis, children short stories, verbs, and the previous studies. A. Corpus Based Analysis in Children Short Stories In these last five decades the work that takes the concept of using corpus has been increased. Corpus, the plural forms are certainly called as corpora, refers to the collection of text, written or spoken, which is systematically gathered. A corpus can also be defined as a broad, principled set of naturally occurring examples of electronically stored languages (Bennet, 2010, p. 2). For such studies, corpus typically refers to a set of authentic machine-readable text that has been selected to describe or represent a condition or variety of a language (Grigaliuniene, 2013, p. 9). Likewise, Lindquist (2009) also believed that corpus is related to electronic. He claimed that corpus is a collection of texts stored on some kind of digital medium to be used by linguists for research purposes or by lexicographers in the production of dictionaries (Lindquist, 2009, p. 3). Nowadays, the word 'corpus' is almost often associated with the term 'electronic corpus,' which is a collection of texts stored on some kind of digital medium to be used by linguists for research purposes or by lexicographers for dictionaries. McCarthy (2004) also described corpus as a collection of written or spoken texts, typically stored in a computer database. We may infer from the above argument that computer has a crucial function in corpus (McCarthy, 2004, p. 1). In this regard, computers and software programs have allowed researchers to fairly quickly and cheaply capture, store and handle vast quantities of data.
    [Show full text]
  • 1. Introduction
    This is the accepted manuscript of the chapter MacLeod, N, and Wright, D. (2020). Forensic Linguistics. In S. Adolphs and D. Knight (eds). Routledge Handbook of English Language and Digital Humanities. London: Routledge, pp. 360-377. Chapter 19 Forensic Linguistics 1. INTRODUCTION One area of applied linguistics in which there has been increasing trends in both (i) utilising technology to assist in the analysis of text and (ii) scrutinising digital data through the lens of traditional linguistic and discursive analytical methods, is that of forensic linguistics. Broadly defined, forensic linguistics is an application of linguistic theory and method to any point at which there is an interface between language and the law. The field is popularly viewed as comprising three main elements: (i) the (written) language of the law, (ii) the language of (spoken) legal processes, and (iii) language analysis as evidence or as an investigative tool. The intersection between digital approaches to language analysis and forensic linguistics discussed in this chapter resides in element (iii), the use of linguistic analysis as evidence or to assist in investigations. Forensic linguists might take instructions from police officers to provide assistance with criminal investigations, or from solicitors for either side preparing a prosecution or defence case in advance of a criminal trial. Alternatively, they may undertake work for parties involved in civil legal disputes. Forensic linguists often appear in court to provide their expert testimony as evidence for the side by which they were retained, though it should be kept in mind that standards here are required to be much higher than they are within investigatory enterprises.
    [Show full text]
  • Distributed Memory Bound Word Counting for Large Corpora
    Democratic and Popular Republic of Algeria Ministry of Higher Education and Scientific Research Ahmed Draia University - Adrar Faculty of Science and Technology Department of Mathematics and Computer Science A Thesis Presented to Fulfil the Master’s Degree in Computer Science Option: Intelligent Systems. Title: Distributed Memory Bound Word Counting For Large Corpora Prepared by: Bekraoui Mohamed Lamine & Sennoussi Fayssal Taqiy Eddine Supervised by: Mr. Mediani Mohammed In front of President : CHOUGOEUR Djilali Examiner : OMARI Mohammed Examiner : BENATIALLAH Djelloul Academic Year 2017/2018 Abstract: Statistical Natural Language Processing (NLP) has seen tremendous success over the recent years and its applications can be met in a wide range of areas. NLP tasks make the core of very popular services such as Google translation, recommendation systems of big commercial companies such Amazon, and even in the voice recognizers of the mobile world. Nowadays, most of the NLP applications are data-based. Language data is used to estimate statistical models, which are then used in making predictions about new data which was probably never seen. In its simplest form, computing any statistical model will rely on the fundamental task of counting the small units constituting the data. With the expansion of the Internet and its intrusion in all aspects of human life, the textual corpora became available in very large amounts. This high availability is very advantageous performance-wise, as it enlarges the coverage and makes the model more robust both to noise and unseen examples. On the other hand, training systems on large data quantities raises a new challenge to the hardware resources, as it is very likely that the model will not fit into main memory.
    [Show full text]
  • Unit 3: Available Corpora and Software
    Corpus building and investigation for the Humanities: An on-line information pack about corpus investigation techniques for the Humanities Unit 3: Available corpora and software Irina Dahlmann, University of Nottingham 3.1 Commonly-used reference corpora and how to find them This section provides an overview of commonly-used and readily available corpora. It is also intended as a summary only and is far from exhaustive, but should prove useful as a starting point to see what kinds of corpora are available. The Corpora here are divided into the following categories: • Corpora of General English • Monitor corpora • Corpora of Spoken English • Corpora of Academic English • Corpora of Professional English • Corpora of Learner English (First and Second Language Acquisition) • Historical (Diachronic) Corpora of English • Corpora in other languages • Parallel Corpora/Multilingual Corpora Each entry contains the name of the corpus and a hyperlink where further information is available. All the information was accurate at the time of writing but the information is subject to change and further web searches may be required. Corpora of General English The American National Corpus http://www.americannationalcorpus.org/ Size: The first release contains 11.5 million words. The final release will contain 100 million words. Content: Written and Spoken American English. Access/Cost: The second release is available from the Linguistic Data Consortium (http://projects.ldc.upenn.edu/ANC/) for $75. The British National Corpus http://www.natcorp.ox.ac.uk/ Size: 100 million words. Content: Written (90%) and Spoken (10%) British English. Access/Cost: The BNC World Edition is available as both a CD-ROM or via online subscription.
    [Show full text]
  • Lexicography Between NLP and Linguistics: Aspects of Theory and Practice
    LEXICOGRAPHY IN GLOBAL CONTEXTS 25 Lexicography between NLP and Linguistics: Aspects of Theory and Practice Lars Trap-Jensen Society for Danish Language and Literature E-mail: [email protected] Abstract Over the last hundred years, lexicography has witnessed three major revolutions: a descriptive revolution at the turn of the 20th century, a corpus revolution in the second half of the 20th century, and the digital revolution which is happening right now. Finding ourselves in the middle of a radical change, most of us have difficulties orient- ing ourselves and knowing where all this is leading. I don’t pretend to know the answers but one thing is clear: we cannot ignore it and carry on as normal. In this article, I will discuss how lexicography and natural language processing can mutually benefit from each other and how lexicography could meet some of the needs that NLP has. I suggest that lexicographers shift their focus from a single dictionary towards the lexical database behind it. Keywords: e-lexicography, NLP, interoperability, data structure 1 Introduction Let me start with a confession: In my career, I have never been much concerned with natural language processing from a theoretical point of view. The reason is simple: My main interest lies with how natural language works, and I find that formal models of language have so far been unable to come up with con- vincing explanations of the way language works. This applies both to formal linguistic theories such as generative grammar in the Chomskyan tradition and to computational linguistic models such as the ones used in NLP.
    [Show full text]
  • Types of Corpora and Some Famous (English) Examples
    Types of corpora and some famous (English) examples Balanced, representative Texts selected in pre-defined proportions to mirror a particular language or language variety. Examples: Brown family <http://khnt.aksis.uib.no/icame/manuals/> – Written. 1 m words, 15 text categories, 500 texts a 2,000 words • American English: Brown (1961) Frown (1992) • British English: LOB (1961), FLob (1991) • Indian (Kolhapur), NZ (Wellington), Australian (ACE) BNC: British National Corpus <http://www.natcorp.ox.ac.uk> – 100 m words, 10% spoken – Carefully composed to be ‘balanced’ – Oxford access at <https://ota.oerc.ox.ac.uk//bncweb-cgi/BNCquery.pl> Monitor New texts added by and by to ‘monitor’ language change. Examples: BoE: Bank of English <http://www.collins.co.uk/books.aspx?group=153> – Written and spokne, much newspaper/media language – Different varieties and text categories – Part can be searched online COCA: Corpus of Contemporary American English <http://www.americancorpus.org/> – currently 385 mwd – 5 genres (one spo) a 4 mwd from 1990 – 2008 – Searchable online Parallel (translation) Same texts in two (or more) languages Examples: OPUS open source parallel corpus <http://urd.let.rug.nl/tiedeman/OPUS/ > – Access to aligned corpora, mainly EU texts. – Unknown size. ENPC English-Norwegian Parallel Corpus <http://www.hf.uio.no/ilos/forskning/forskningsprosjekter/enpc/> – Originally English and Norwegian originals with Norwegian and English translations, now also German, Dutch, Portuguese. – 50 text extracts in each direction, fiction and non-fiction
    [Show full text]
  • Intégration De Verbnet Dans Un Réalisateur Profond
    Université de Montréal Intégration de VerbNet dans un réalisateur profond par Daniel Galarreta-Piquette Département de linguistique et de traduction Faculté des arts et des sciences Mémoire présenté en vue de l’obtention du grade de Maître ès arts (M.A.) en linguistique Août 2018 c Daniel Galarreta-Piquette, 2018 RÉSUMÉ La génération automatique de texte (GAT) a comme objectif de produire du texte com- préhensible en langue naturelle à partir de données non-linguistiques. Les générateurs font es- sentiellement deux tâches : d’abord ils déterminent le contenu d’un message à communiquer, puis ils sélectionnent les mots et les constructions syntaxiques qui serviront à transmettre le message, aussi appellée la réalisation linguistique. Pour générer des textes aussi naturels que possible, un système de GAT doit être doté de ressources lexicales riches. Si on veut avoir un maximum de flexibilité dans les réalisations, il nous faut avoir accès aux différentes pro- priétés de combinatoire des unités lexicales d’une langue donnée. Puisque les verbes sont au cœur de chaque énoncé et qu’ils contrôlent généralement la structure de la phrase, il faudrait encoder leurs propriétés afin de produire du texte exploitant toute la richesse des langues. De plus, les verbes ont des propriétés de combinatoires imprévisibles, c’est pourquoi il faut les encoder dans un dictionnaire. Ce mémoire porte sur l’intégration de VerbNet, un dictionnaire riche de verbes de l’anglais et de leurs comportements syntaxiques, à un réalisateur profond, GenDR. Pour procéder à cette implémentation, nous avons utilisé le langage de programmation Python pour extraire les données de VerbNet et les manipuler pour les adapter à GenDR, un réalisateur profond basé sur la théorie Sens-Texte.
    [Show full text]
  • The Spoken British National Corpus 2014
    The Spoken British National Corpus 2014 Design, compilation and analysis ROBBIE LOVE ESRC Centre for Corpus Approaches to Social Science Department of Linguistics and English Language Lancaster University A thesis submitted to Lancaster University for the degree of Doctor of Philosophy in Linguistics September 2017 Contents Contents ............................................................................................................. i Abstract ............................................................................................................. v Acknowledgements ......................................................................................... vi List of tables .................................................................................................. viii List of figures .................................................................................................... x 1 Introduction ............................................................................................... 1 1.1 Overview .................................................................................................................................... 1 1.2 Research aims & structure ....................................................................................................... 3 2 Literature review ........................................................................................ 6 2.1 Introduction ..............................................................................................................................
    [Show full text]
  • The BT Correspondence Corpus 1853- 1982: the Development and Analysis of Archival Language Resources
    The BT Correspondence Corpus 1853- 1982: The development and analysis of archival language resources Morton, R. Submitted version deposited in Coventry University’s Institutional Repository Original citation: Morton, R. (2016) The BT Correspondence Corpus 1853-1982: The development and analysis of archival language resources. Unpublished PhD Thesis. Coventry: Coventry University Copyright © and Moral Rights are retained by the author. A copy can be downloaded for personal non-commercial research or study, without prior permission or charge. This item cannot be reproduced or quoted extensively from without first obtaining permission in writing from the copyright holder(s). The content must not be changed in any way or sold commercially in any format or medium without the formal permission of the copyright holders. Some materials have been removed from this thesis due to third party. Pages where material has been removed are clearly marked in the electronic version. The unabridged version of the thesis can be viewed at the Lanchester Library, Coventry University. The BT Correspondence Corpus 1853-1982: The development and analysis of archival language resources RALPH MORTON Thesis presented to the Graduate School of Coventry University in partial fulfilment of the requirements for the degree of Doctor of Philosophy in CORPUS AND HISTORICAL LINGUISTICS COVENTRY UNIVERSITY MAY 2016 ACKNOWLEDGMENTS This research couldn’t have happened without the support of lots of people. Firstly I’d like to thank the New Connections project partners at The National Archives and the BT Archives. Siân Wynn-Jones and David Hay at BT in particular have been very helpful in providing information about their collections and archives in general.
    [Show full text]
  • The Corpus of Contemporary American English As the First Reliable Monitor Corpus of English
    The Corpus of Contemporary American English as the first reliable monitor corpus of English ............................................................................................................................................................ Mark Davies Brigham Young University, Provo, UT, USA ....................................................................................................................................... Abstract The Corpus of Contemporary American English is the first large, genre-balanced corpus of any language, which has been designed and constructed from the ground up as a ‘monitor corpus’, and which can be used to accurately track and study recent changes in the language. The 400 million words corpus is evenly divided between spoken, fiction, popular magazines, newspapers, and academic journals. Most importantly, the genre balance stays almost exactly the same from year to year, which allows it to accurately model changes in the ‘real world’. After discussing the corpus design, we provide a number of concrete examples of how the corpus can be used to look at recent changes in English, including morph- ology (new suffixes –friendly and –gate), syntax (including prescriptive rules, Correspondence: quotative like, so not ADJ, the get passive, resultatives, and verb complementa- Mark Davies, Linguistics and tion), semantics (such as changes in meaning with web, green, or gay), and lexis–– English Language, Brigham including word and phrase frequency by year, and using the corpus architecture Young University.
    [Show full text]
  • An Analysis of Factive Cognitive Verb Complementation Patterns Used by Elt Students
    AN ANALYSIS OF FACTIVE COGNITIVE VERB COMPLEMENTATION PATTERNS USED BY ELT STUDENTS Doktora Tezi Serap ATASEVER BELLİ Eskişehir 2019 AN ANALYSIS OF FACTIVE COGNITIVE VERB COMPLEMENTATION PATTERNS USED BY ELT STUDENTS Serap ATASEVER BELLİ PhD DISSERTATION English Language Teaching Program Department of Foreign Language Education Supervisor: Prof. Dr. İlknur KEÇİK Eskişehir Anadolu University Graduate School of Educational Sciences May 2019 This doctoral dissertation study is a part of a larger project and has been funded by The Scientific and Technological Research Council of Turkey (TÜBİTAK) Social and Human Sciences Research Grant Group (SOBAG- 1001) with the Project No. 118K130. ÖZET İNGİLİZCE ÖĞRETMENLİĞİ PROGRAMI’NDA OKUYAN ÖĞRENCİLERİN KULLANDIKLARI OLGUSAL BİLİŞ EYLEMLERİNİN YANULAMLAMA GÖRÜNÜMLERİNİN İNCELENMESİ Serap ATASEVER BELLİ Yabancı Diller Eğitimi Anabilim Dalı Anadolu Üniversitesi, Eğitim Bilimleri Enstitüsü, Mayıs 2019 Danışman: Prof. Dr. İlknur KEÇİK Bu araştırmanın amacı, İngilizce’yi yabancı dil olarak olarak öğrenen Türk öğrencilerin olgusal biliş eylemlerini (bil-, anla-, hatırla-, unut-, ve pişman ol-) yanulamlama (eylemin seçtiği üyeler) özellikleri açısından anlama ve kullanım yeterliliklerini ortaya çıkarmaktır. Bu amaç doğrultusunda, gömülü karma yöntem benimsenmiş, hem nitel hem nicel veri toplama yöntemleri kullanılmıştır. Araştırmanın katılımcılarını bir devlet üniversitesinde İngilizce Öğretmenliği Bölümü’nde okuyan 269 öğrenci oluşturmaktadır. Veri toplama aracı olarak dört çeşit test (Cümle Tamamlama,
    [Show full text]