Towards Better Context-Aware Lexical Semantics: Adjusting Contextualized Representations Through Static Anchors

Total Page:16

File Type:pdf, Size:1020Kb

Towards Better Context-Aware Lexical Semantics: Adjusting Contextualized Representations Through Static Anchors Towards Better Context-aware Lexical Semantics: Adjusting Contextualized Representations through Static Anchors Qianchu Liu, Diana McCarthy, Anna Korhonen Language Technology Lab, TAL, University of Cambridge, UK [email protected], [email protected], [email protected] Abstract capturing context-sensitive word meanings, static embeddings still achieve better performance than One of the most powerful features of con- textualized models is their dynamic embed- contextualized embeddings in traditional context- dings for words in context, leading to state-of- independent lexical semantic tasks including word the-art representations for context-aware lex- similarity and analogy (Wang et al., 2019). This ical semantics. In this paper, we present a suggests that static embeddings have the potential post-processing technique that enhances these to offer complementary semantic information to en- representations by learning a transformation hance contextualized models for lexical semantics. through static anchors. Our method requires We bridge the aforementioned gaps and propose only another pre-trained model and no labeled a general framework that improves contextualized data is needed. We show consistent improve- ment in a range of benchmark tasks that test representations by leveraging other pre-trained con- contextual variations of meaning both across textualized/static models. We achieve this by using different usages of a word and across different static anchors (the average contextual representa- words as they are used in context. We demon- tions for each word) to transform the original con- strate that while the original contextual rep- textualized model, guided by the embedding space resentations can be improved by another em- from another model. We assess the overall quality bedding space from either contextualized or of a model’s lexical semantic representation by two static models, the static embeddings, which have lower computational requirements, pro- Inter Word tasks that measure relations between dif- vide the most gains. ferent words in context. We also evaluate on three Within Word tasks that test the contextual effect 1 Introduction from different usages of the same word/word pair. Word representations are fundamental in Natural Our method obtains consistent improvement across Language Processing (NLP) (Bengio et al., 2003). all these context-aware lexical semantic tasks. We Recently, there has been a surge of contextualized demonstrate the particular strength of leveraging models that achieve state-of-the-art in many NLP static embeddings, and offer insights on the reasons benchmark tasks (Peters et al., 2018; Devlin et al., behind the improvement. Our method also has min- 2019; Liu et al., 2019b; Yang et al., 2019). Even imum computational complexity and requires no better performance has been reported from fine- labeled data. tuning or training multiple contextualized models 2 Background for a specific task such as question answering (De- vlin et al., 2019; Xu et al., 2020). However, little This section briefly introduces the contextual- has been explored on directly leveraging the many ized/static models that we experimented in this off-the-shelf pre-trained models to improve task- study. For static models, we select three repre- independent representations for lexical semantics. sentative methods. SGNS (Mikolov et al., 2013), Furthermore, classic static embeddings are often as the most successful variant of word2vec, trains overlooked in this trend towards contextualized a log linear model to predict context words given models. As opposed to contextualized embeddings a target word with negative sampling in the objec- that generate dynamic representations for words tive. FastText improves over SGNS by training in context, static embeddings such as word2vec at the n-gram level and can generalize to unseen (Mikolov et al., 2013) assign one fixed representa- words (Bojanowski et al., 2017). In addition to tion for each word. Despite being less effective in these two prediction-based models, we also include 4066 Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, pages 4066–4075, November 16–20, 2020. c 2020 Association for Computational Linguistics one count-based model, GloVe (Pennington et al., as they provide one fixed representation per word, 2014). GloVe is trained to encode semantic rela- and therefore correspond to word embeddings from tions exhibited in the ratios of word-word occur- a static model such as FastText. We also use these rence probabilities into word vectors. anchors to align between contextualized models. As opposed to static embeddings, contextualized To form the vocabulary for creating static anchors models provide dynamic lexical representations as in our experiments, we take the top 200k most fre- hidden layers in deep neural networks typically quent words and extract their contexts from English pre-trained with language modeling objectives. In Wikipedia. our study, we choose three state-of-the-art con- To describe the method in more detail, we repre- textualized models. BERT (Devlin et al., 2019) sent the anchor embeddings from the original con- trains bidirectional transformers (Vaswani et al., textualized model as our source matrix S, and the 2017) with masked language modeling and next corresponding representations from another con- sentence prediction objectives. Liu et al.(2019b)’s textualized/static model as target matrix T. si and RoBERTa further improves upon BERT by care- ti are the source and target vectors for the ith word fully optimizing a series of design decisions. XL- in the vocabulary (V ). We first find an orthogonal Net (Yang et al., 2019) takes a generalized auto- alignment matrix W that rotates the target space to regressive pre-training approach and integrates the source space by solving the least squares linear ideas from Transformer-XL (Dai et al., 2019). For regression problem in Eq.1. W is found through the best performance, we use the Large Cased1 Procrustes analysis (Schonemann¨ , 1966). variant for each contextualized model. Since our jV j study focuses on generic lexical representations X 2 T W = arg min kWti − sik s:t: W W = I: (1) W and many of the lexical semantic tasks do not i=1 provide training data, we extract features2 from these contextualized models without fine-tuning As described in Eq.2, we then learn a linear the weights for a specific task. This feature-based mapping M to transform the source space towards approach is also more efficient compared with fine- the average of source and the rotated target space, tuning the increasingly larger models which can by minimizing the squared Euclidean distance be- have hundreds of millions of parameters. tween each transformed source vector Msi and the mean vector µi (µi = (si + Wti)=2). M is the 3 Method mapping we will use to transform the original con- textualized space. Following Doval et al.(2018), Our method3 is built from a recently proposed M is found via a closed-form solution. cross-lingual alignment technique called meeting jV j in the middle X 2 (Doval et al., 2018). Their method M = arg min kMsi − µik (2) M relies on manual translations to learn a transfor- i mation over an orthogonal alignment for better For improved alignment quality, as advised by cross-lingual static embeddings. We show that Artetxe et al.(2016), we normalize and mean- by a similar alignment + transformation technique, center4 the embeddings in S and T a priori. we can improve monolingual contextualized em- beddings without resorting to any labeled data. 4 Experiments The direct correspondence among contextual- Task Descriptions5 We evaluate on three Within ized and static embeddings for alignment is not Word tasks. Usage Similarity (Usim)(Erk et al., straightforward, as contextualized models can com- 2013) dataset measures graded similarity of the pute infinite representations for infinite contexts. same word in pairs of different contexts on the Inspired by previous study (Schuster et al., 2019) scale from 1 to 5. Word in Context (WiC)(Pilehvar that found contextualized embeddings roughly and Camacho-Collados, 2019) dataset challenges a form word clusters, we take the average of each system to predict a binary choice of whether a pair word’s contextual representations as anchors of a of contexts for the same word belongs to the same contextualized model. We call them static anchors 4We pre-process representations with the same centering 11024 dimensions with case-preserving vocabulary. and normalization in all tasks. Our reported results are similar 2AppendixA contains more details on feature extraction. or better than the results from un-preprocessed representations. 3Implementation details are listed in AppendixB. 5AppendixC reports details for each task and experiment. 4067 meaning or not. We follow the advised training word relations (Wang et al., 2019) into a contextu- scheme in the original paper to learn a cosine simi- alized model. At the same time, static embeddings larity threshold on the representations. The recently consistently improve performance in Within Word proposed CoSimlex (Armendariz et al., 2019) task tasks in 24 out of the total 27 configurations, reas- provides contexts for selected word pairs from the suring us that the contextualization power of the word similarity benchmark SimLex-999 (Hill et al., original contextual space is not only preserved but 2015) and measures the graded contextual effect. even enhanced. Overall, FastText is the most robust We use the English
Recommended publications
  • Logophoricity in Finnish
    Open Linguistics 2018; 4: 630–656 Research Article Elsi Kaiser* Effects of perspective-taking on pronominal reference to humans and animals: Logophoricity in Finnish https://doi.org/10.1515/opli-2018-0031 Received December 19, 2017; accepted August 28, 2018 Abstract: This paper investigates the logophoric pronoun system of Finnish, with a focus on reference to animals, to further our understanding of the linguistic representation of non-human animals, how perspective-taking is signaled linguistically, and how this relates to features such as [+/-HUMAN]. In contexts where animals are grammatically [-HUMAN] but conceptualized as the perspectival center (whose thoughts, speech or mental state is being reported), can they be referred to with logophoric pronouns? Colloquial Finnish is claimed to have a logophoric pronoun which has the same form as the human-referring pronoun of standard Finnish, hän (she/he). This allows us to test whether a pronoun that may at first blush seem featurally specified to seek [+HUMAN] referents can be used for [-HUMAN] referents when they are logophoric. I used corpus data to compare the claim that hän is logophoric in both standard and colloquial Finnish vs. the claim that the two registers have different logophoric systems. I argue for a unified system where hän is logophoric in both registers, and moreover can be used for logophoric [-HUMAN] referents in both colloquial and standard Finnish. Thus, on its logophoric use, hän does not require its referent to be [+HUMAN]. Keywords: Finnish, logophoric pronouns, logophoricity, anti-logophoricity, animacy, non-human animals, perspective-taking, corpus 1 Introduction A key aspect of being human is our ability to think and reason about our own mental states as well as those of others, and to recognize that others’ perspectives, knowledge or mental states are distinct from our own, an ability known as Theory of Mind (term due to Premack & Woodruff 1978).
    [Show full text]
  • Context Pragmatics Definition of Pragmatics
    Pragmatics • to ask a question: Maybe Sandy’s reassuring you that Kim’ll get home okay, even though she’s walking home late at night. Sandy: She’s got more than lipstick and Kleenex in that purse of hers. • What is Pragmatics? You: Kim’s got a knife? • Context and Why It’s Important • Speech Acts – Direct Speech Acts – Indirect Speech Acts • How To Make Sense of Conversations – Cooperative Principle – Conversational Maxims Linguistics 201, Detmar Meurers Handout 3 (April 9, 2004) 1 3 Definition of Pragmatics Context Pragmatics is the study of how language is used and how language is integrated in context. What exactly are the factors which are relevant for an account of how people use language? What is linguistic context? Why must we consider context? We distinguish several types of contextual information: (1) Kim’s got a knife 1. Physical context – this encompasses what is physically present around the speakers/hearers at the time of communication. What objects are visible, where Sentence (1) can be used to accomplish different things in different contexts: the communication is taking place, what is going on around, etc. • to make an assertion: You’re sitting on a beach, thinking about how to open a coconut, when someone (2) a. Iwantthat book. observes “Kim’s got a knife”. (accompanied by pointing) b. Be here at 9:00 tonight. • to give a warning: (place/time reference) Kim’s trying to bully you and Sandy into giving her your lunch money, and Sandy just turns around and starts to walk away. She doesn’t see Kim bring out the butcher knife, and hears you yell behind her, “Kim’s got a knife!” 2 4 2.
    [Show full text]
  • Arxiv:2106.08037V1 [Cs.CL] 15 Jun 2021 Alternative Ways the World Could Be
    The Possible, the Plausible, and the Desirable: Event-Based Modality Detection for Language Processing Valentina Pyatkin∗ Shoval Sadde∗ Aynat Rubinstein Bar Ilan University Bar Ilan University Hebrew University of Jerusalem [email protected] [email protected] [email protected] Paul Portner Reut Tsarfaty Georgetown University Bar Ilan University [email protected] [email protected] Abstract (1) a. We presented a paper at ACL’19. Modality is the linguistic ability to describe b. We did not present a paper at ACL’20. events with added information such as how de- sirable, plausible, or feasible they are. Modal- The propositional content p =“present a paper at ity is important for many NLP downstream ACL’X” can be easily verified for sentences (1a)- tasks such as the detection of hedging, uncer- (1b) by looking up the proceedings of the confer- tainty, speculation, and more. Previous studies ence to (dis)prove the existence of the relevant pub- that address modality detection in NLP often p restrict modal expressions to a closed syntac- lication. The same proposition is still referred to tic class, and the modal sense labels are vastly in sentences (2a)–(2d), but now in each one, p is different across different studies, lacking an ac- described from a different perspective: cepted standard. Furthermore, these senses are often analyzed independently of the events that (2) a. We aim to present a paper at ACL’21. they modify. This work builds on the theoreti- b. We want to present a paper at ACL’21. cal foundations of the Georgetown Gradable Modal Expressions (GME) work by Rubin- c.
    [Show full text]
  • Semantics and Context-Dependence: Towards a Strawsonian Account∗
    Semantics and Context-Dependence: Towards a Strawsonian Account∗ Richard G Heck, Jr Brown University Some time in the mid-1990s, I began to encounter variations on the following argument, which was popular with certain of my colleagues and with students who had fallen under their influence. (i) Semantics is about truth-conditions. (ii) Only utterances have truth-conditions. (iii) Hence, a semantic theory must assign truth-conditions to utterances. (iv) Only a theory that incorporated a complete theory of human rationality could assign truth-conditions to utterances. (v) So there is no such thing as a semantic theory (in anything like the usual sense). The argument need not be formulated in terms of truth-conditions. The first premise could equally be stated as: Semantics is about the expression of propositions. The rest of the argument then adapts smoothly. So (i) is meant to be obvious and uncontroversial. With regard to the second premise, it is important to appreciate the force of the word “only”: (ii) asserts that (perhaps with a very few exceptions, such as statements of pure mathematics) sentences never have truth-conditions; only utterances do. So what underwrites (ii) is not just the existence of context-dependence but, rather, its ubiquity. As has become clear over the last few decades, it is not just the obvious expressions—like “I”, “here”, “this”, and the like—whose meaning seems to vary with the circumstances of utterance. All of the following sentences seem as if they could express different propositions on different occasions, due to context-dependence connected with the italicized expressions: • No-one passed the test.
    [Show full text]
  • Lexical Semantics Ling571 Deep Processing Techniques for NLP February 23, 2015 What Is a Plant?
    Lexical Semantics Ling571 Deep Processing Techniques for NLP February 23, 2015 What is a plant? There are more kinds of plants and animals in the rainforests than anywhere else on Earth. Over half of the millions of known species of plants and animals live in the rainforest. Many are found nowhere else. There are even plants and animals in the rainforest that we have not yet discovered. The Paulus company was founded in 1938. Since those days the product range has been the subject of constant expansions and is brought up continuously to correspond with the state of the art. We’re engineering, manufacturing, and commissioning world-wide ready-to-run plants packed with our comprehensive know-how. Lexical Semantics So far, word meanings discrete Constants, predicates, functions Lexical Semantics So far, word meanings discrete Constants, predicates, functions Focus on word meanings: Relations of meaning among words Similarities & differences of meaning in sim context Lexical Semantics So far, word meanings discrete Constants, predicates, functions Focus on word meanings: Relations of meaning among words Similarities & differences of meaning in sim context Internal meaning structure of words Basic internal units combine for meaning Terminology Lexeme: Form: Orthographic/phonological + meaning Terminology Lexeme: Form: Orthographic/phonological + meaning Represented by lemma Lemma: citation form; infinitive in inflection Sing: sing, sings, sang, sung,… Terminology Lexeme: Form: Orthographic/phonological + meaning Represented by lemma Lemma: citation form; infinitive in inflection Sing: sing, sings, sang, sung,… Lexicon: finite list of lexemes Sources of Confusion Homonymy: Words have same form but different meanings Generally same POS, but unrelated meaning Sources of Confusion Homonymy: Words have same form but different meanings Generally same POS, but unrelated meaning E.g.
    [Show full text]
  • Person, Context and Perspective
    To appear in Rivista di Linguistica – Italian Journal of Linguistics in 2005. Person, Context and Perspective Ken Safir, Rutgers University ABSTRACT: It is argued that the indexicality of first person pronouns is arises from a restriction on the pronouns themselves, as opposed to any operator that binds them. The nature of this restriction is an asyntactic constant function that picks out individuals to the context of utterance (following Kaplan, 1989)). Constant function pronouns do not require an antecedent, neither an operator nor an argument, although this does not prevent them from participating in bound readings if an appropriate antecedent is introduced. The notion that agents of contexts and agents of propositional attitudes are versions of the same operator-variable relation is thus rejected, along with certain less fine-grained versions of the nature of de se interpretation. Consequently, indexical pronouns such as first person ones contrast with logophoric pronouns, which are necessarily operator-bound by perspectival operators introduced by propositional attitude verbs. Scope-sensitive properties of operator-binding and the perspectival interpretations that are imposed on logophoricity distinguish the latter from constant function phenomena, which are sensitive neither to scope, as it is usually treated, nor perspectival shifts. Constant function phenomena are also detectable as restrictions on third person forms, and two such examples are lightly explored: the English generic pronoun one and the proximate/obviative distinction in languages like Fox. 1.0 Introduction In this essay I defend the thesis that the indexicality of first person pronouns is a restriction on the pronouns themselves, as opposed to any operator that binds them.
    [Show full text]
  • Pragmatics Context Deixis
    Llinguistics 3rd Year College of Education for Women / Tikrit University English Department Lecturer: Dr. Israa Burhanuddin Abdurrahman Pragmatics The study of what speakers mean, or 'speaker meaning', is called Pragmatics. Pragmatics is the study of invisible meaning or how we recognize what is meant even when it is not actually said. Speakers depend on a lot of shared assumptions and expectations. You use the meanings of the words, in combination, and the context in which they occur, and you try to arrive at what the writer of the sign intended his message to convey Context • We have got two kinds of contexts. 1- one kind is best described as linguistic context, also known as cotext. • The co-text of a word is the set of other words used in the same phrase or sentence. • e.g. I get to the bank to cash a cheque. • Bank is homonym. By looking at other words in the sentence we know which type of bank is intended. 2- another type of context is described as physical context . Our • understanding of what we read and hear is tied to the physical context, particularly the time and place. • e.g. The word bank on the wall of a building in a city. Deixis • There are some words in the language that cannot be interpreted at all, unless the physical context is known. “here, there, this, that, now, then, yesterday, come” , pronouns, such as “I, you, him, her, them”, e.g. You will have to bring that back tomorrow, because they are not here now. This sentence is vague.
    [Show full text]
  • Indexicality and Context-Shift François Recanati
    Indexicality and context-shift François Recanati To cite this version: François Recanati. Indexicality and context-shift. Workshop on indexicals, speech acts and logophors, Nov 2004, Harvard University. ijn_00000552 HAL Id: ijn_00000552 https://jeannicod.ccsd.cnrs.fr/ijn_00000552 Submitted on 29 Nov 2004 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Indexicality and context-shift François Recanati Institut Jean Nicod/Harvard University Workshop on Indexicality, Speech Acts and Logophors Harvard University, November 20th, 2004 1. Preliminaries 1.1 Indexicality and semantic under-specification Indexicals are expressions whose semantic value systematically depends upon the context of utterance, and whose linguistic meaning somehow encodes this dependency upon the context of utterance. Thus I do not count as indexical in the strict sense those expressions whose semantic value depends upon the context merely because they are semantically under- specified and function as free variables to which a value must be contextually assigned. Whenever an expression
    [Show full text]
  • Context Effects and Lexical Ambiguity Processing: an Activation-Based Account Frazier Orgain
    University of Richmond UR Scholarship Repository Master's Theses Student Research 5-2002 Context effects and lexical ambiguity processing: an activation-based account Frazier Orgain Follow this and additional works at: http://scholarship.richmond.edu/masters-theses Recommended Citation Orgain, Frazier, "Context effects and lexical ambiguity processing: an activation-based account" (2002). Master's Theses. Paper 810. This Thesis is brought to you for free and open access by the Student Research at UR Scholarship Repository. It has been accepted for inclusion in Master's Theses by an authorized administrator of UR Scholarship Repository. For more information, please contact [email protected]. Project Name: ~~\V\,__ f yo..,:t;t/( Z,OO "Z- Date: Patron: Specialist: ~/l'l//S \)\ \) ~iL Project Description: N~tt,r Tke!~ Hardware Specs: CONTEXT EFFECTS AND LEXICAL AMBIGUITY PROCESSING: AN ACTIVATION-BASED ACCOUNT By Frazer Orgain B.S., Virginia Military Institute, 1999 A Thesis Submitted to the Graduate Faculty of the University of Richmond in Candidacy for the degree of MASTER OF ARTS m Psychology May, 2002 University of Richmond, Virginia I certify that I have read this thesis and find that, in scope and quality, it satisfies the requirements for the degree of Master of Arts. Ping Li, Thesis Advisor Jane M.~ ~ &wi Beth Crawford · 11 Abstract Many studies have been conducted to test the effects of ambiguous words in sentence processing. There are two views: the modularity hypothesis and the interactive hypothesis that dominate this field of study. The effect of ambiguity has been tested in many ways, including gating, cross-modal priming, naming, and self-paced reading.
    [Show full text]
  • Knowledge Frames and Speech Act Comprehension * Teun A. Van Dijk
    CONTEXT AND COGNITION: KNOWLEDGE FRAMES AND SPEECH ACT COMPREHENSION * TEUN A. VAN DIJK This paper is about the cognitive foundations of pragmatic theories. Besides the fact that the usual appropriateness conditions for speech acts, which are given in cognitive terms, such as S knows/believes/wants ... (that) p , require empirical investigation, a sound theory of prag- matics must also explain how certain utterances in certain contexts are actually understood as certain speech acts. Speech act comprehension is based on rules and strategies for so-called context analysis , in which (epistemic) frames play an important role in the analysis of social context, social frames, and interaction type. Results of context analysis are then matched with those of pragmatic sentence analysis, viz. the illocutionary act indicating devices. Finally, some results from the cognitive analysis of discourse processing are applied in a brief account of the comprehension of speech act sequences and macro-speech acts. 1. The foundations of pragmatics The philosophical and linguistic study of pragmatics requires an analysis of its foundations. This basis of pragmatic theories is on the one hand conceptual, e.g. in the analysis of action and interaction, and on the other hand empirical, viz. in the investigation of the psychological and social properties of language processing in communicative interaction. One of the crucial components in an integrated theory of communicative inter- action by natural language is constituted by a cognitive theory of language use. Such a cognitive theory will not only provide insight into the processes and structures involved in the actual production, comprehension, storage, reproduction * This paper is intended as an introductory survey of some problems in the interdisciplinary domain of pragmatics and cognitive psychology.
    [Show full text]
  • Long-Distance Reflexivization and Logophoricity in the Dargin Language Muminat Kerimova Florida International University
    Florida International University FIU Digital Commons MA in Linguistics Final Projects College of Arts, Sciences & Education 2017 Long-Distance Reflexivization and Logophoricity in the Dargin Language Muminat Kerimova Florida International University Follow this and additional works at: https://digitalcommons.fiu.edu/linguistics_ma Part of the Linguistics Commons Recommended Citation Kerimova, Muminat, "Long-Distance Reflexivization and Logophoricity in the Dargin Language" (2017). MA in Linguistics Final Projects. 3. https://digitalcommons.fiu.edu/linguistics_ma/3 This work is brought to you for free and open access by the College of Arts, Sciences & Education at FIU Digital Commons. It has been accepted for inclusion in MA in Linguistics Final Projects by an authorized administrator of FIU Digital Commons. For more information, please contact [email protected]. FLORIDA INTERNATIONAL UNIVERSITY Miami, Florida LONG-DISTANCE REFLEXIVIZATION AND LOGOPHORICITY IN THE DARGIN LANGUAGE A thesis submitted in partial fulfillment of the requirements for the degree of MASTER OF ARTS in LINGUISTICS by Muminat Kerimova 2017 ABSTRACT OF THE THESIS LONG-DISTANCE REFLEXIVIZATION AND LOGOPHORICITY IN THE DARGIN LANGUAGE by Muminat Kerimova Florida International University, 2017 Miami, Florida Professor Ellen Thompson, Major Professor The study of anaphora challenges us to determine the conditions under which the pronouns of a language are associated with possible antecedents. One of the theoretical questions is whether the distribution of pronominal forms is best explained by a syntactic, semantic or discourse level analysis. A more practical question is how we distinguish between anaphoric elements, e.g. what are the borders between the notions of pronouns, locally bound reflexives and long-distance reflexives? The study analyzes the anaphora device saj in Dargin that is traditionally considered to be a long-distance reflexivization language.
    [Show full text]
  • "Context in the Analysis of Discourse and Interaction" In: The
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Open Repository and Bibliography - Luxembourg Context in the Analysis of Discourse and Interaction INGRID DE SAINT-GEORGES Context is a central concept in the analysis of discourse and interaction in all the major research traditions in applied and sociolinguistics. Early linguistics did not display much sensitivity to it (Scollon, 1998, p. 80) and tended to study utterances in isolation and with- out reference to context. Today, however, there seems to be a general consensus around the idea that we understand utterances because they fi t or make sense within particular situations. Studies of discourse and interaction include some orientation to context, if nothing else because “language is always produced by someone to someone else, at a particular time and place, with a purpose and so forth” (Blommaert, 2005, p. 39). Beyond general agreement that “context should be taken into account” (Jones, 2004, p. 22), however, researchers often disagree about what should count as context, or how much context should be taken into account in the analysis. It could be argued that the way context is treated is in fact what usually sets apart and distinguishes different approaches and research traditions (Tracy, 1998). Conversation analysts, for example, often described as interested in identifying universal conversational rules, usually advocate limiting the study of context to those elements which are evoked in the sequential unfolding of the text or the interaction. Institutional context, social background, gender of the participants, and so on therefore should only be attended to by the analyst if they show up in the interaction and are made relevant to the ongoing exchange by the participants.
    [Show full text]