Utterance-Level Multimodal Sentiment Analysis

Total Page:16

File Type:pdf, Size:1020Kb

Utterance-Level Multimodal Sentiment Analysis Utterance-Level Multimodal Sentiment Analysis Veronica´ Perez-Rosas´ and Rada Mihalcea Louis-Philippe Morency Computer Science and Engineering Institute for Creative Technologies University of North Texas University of Southern California [email protected], [email protected] [email protected] Abstract Riloff, 2005; Esuli and Sebastiani, 2006) or large annotated datasets (Maas et al., 2011). Given the During real-life interactions, people are accelerated growth of other media on the Web and naturally gesturing and modulating their elsewhere, which includes massive collections of voice to emphasize specific points or to videos (e.g., YouTube, Vimeo, VideoLectures), im- express their emotions. With the recent ages (e.g., Flickr, Picasa), audio clips (e.g., pod- growth of social websites such as YouTube, casts), the ability to address the identification of Facebook, and Amazon, video reviews are opinions in the presence of diverse modalities is be- emerging as a new source of multimodal coming increasingly important. This has motivated and natural opinions that has been left al- researchers to start exploring multimodal clues for most untapped by automatic opinion anal- the detection of sentiment and emotions in video ysis techniques. This paper presents a content (Morency et al., 2011; Wagner et al., 2011). method for multimodal sentiment classi- In this paper, we explore the addition of speech fication, which can identify the sentiment and visual modalities to text analysis in order to expressed in utterance-level visual datas- identify the sentiment expressed in video reviews. treams. Using a new multimodal dataset Given the non homogeneous nature of full-video consisting of sentiment annotated utter- reviews, which typically include a mixture of posi- ances extracted from video reviews, we tive, negative, and neutral statements, we decided show that multimodal sentiment analysis to perform our experiments and analyses at the ut- can be effectively performed, and that the terance level. This is in line with earlier work on joint use of visual, acoustic, and linguistic text-based sentiment analysis, where it has been modalities can lead to error rate reductions observed that full-document reviews often contain of up to 10.5% as compared to the best both positive and negative comments, which led to performing individual modality. a number of methods addressing opinion analysis at sentence level. Our results show that relying 1 Introduction on the joint use of linguistic, acoustic, and visual Video reviews represent a growing source of con- modalities allows us to better sense the sentiment sumer information that gained increasing interest being expressed as compared to the use of only one from companies, researchers, and consumers. Pop- modality at a time. ular web platforms such as YouTube, Amazon, Another important aspect of this paper is the in- Facebook, and ExpoTV have reported a signifi- troduction of a new multimodal opinion database cant increase in the number of consumer reviews annotated at the utterance level which is, to our in video format over the past five years. Compared knowledge, the first of its kind. In our work, this to traditional text reviews, video reviews provide a dataset enabled a wide range of multimodal senti- more natural experience as they allow the viewer to ment analysis experiments, addressing the relative better sense the reviewer’s emotions, beliefs, and importance of modalities and individual features. intentions through richer channels such as intona- The following section presents related work tions, facial expressions, and body language. in text-based sentiment analysis and audio-visual Much of the work to date on opinion analysis has emotion recognition. Section 3 describes our new focused on textual data, and a number of resources multimodal datasets with utterance-level sentiment have been created including lexicons (Wiebe and annotations. Section 4 presents our multimodal sen- 973 Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics, pages 973–982, Sofia, Bulgaria, August 4-9 2013. c 2013 Association for Computational Linguistics timent analysis approach, including details about such as speech or facial expressions. our linguistic, acoustic, and visual features. Our The only exceptions that we are aware of are the experiments and results on multimodal sentiment findings reported in (Somasundaran et al., 2006; classification are presented in Section 5, with a Raaijmakers et al., 2008; Mairesse et al., 2012; detailed discussion and analysis in Section 6. Metze et al., 2009), where speech and text have been analyzed jointly for the purpose of subjectiv- 2 Related Work ity or sentiment identification, without, however, addressing other modalities such as visual cues; In this section we provide a brief overview of re- and the work reported in (Morency et al., 2011; lated work in text-based sentiment analysis, as well Perez-Rosas et al., 2013), where multimodal cues as audio-visual emotion analysis. have been used for the analysis of sentiment in 2.1 Text-based Subjectivity and Sentiment product reviews, but where the analysis was done Analysis at the much coarser level of full videos rather than individual utterances as we do in our work. The techniques developed so far for subjectivity and sentiment analysis have focused primarily on 2.2 Audio-Visual Emotion Analysis. the processing of text, and consist of either rule- based classifiers that make use of opinion lexicons, Also related to our work is the research done on or data-driven methods that assume the availability emotion analysis. Emotion analysis of speech sig- of a large dataset annotated for polarity. These tools nals aims to identify the emotional or physical and resources have been already used in a large states of a person by analyzing his or her voice number of applications, including expressive text- (Ververidis and Kotropoulos, 2006). Proposed to-speech synthesis (Alm et al., 2005), tracking methods for emotion recognition from speech fo- sentiment timelines in on-line forums and news cus both on what is being said and how is be- (Balog et al., 2006), analysis of political debates ing said, and rely mainly on the analysis of the (Carvalho et al., 2011), question answering (Oh et speech signal by sampling the content at utterance al., 2012), conversation summarization (Carenini et or frame level (Bitouk et al., 2010). Several re- al., 2008), and citation sentiment detection (Athar searchers used prosody (e.g., pitch, speaking rate, and Teufel, 2012). Mel frequency coefficients) for speech-based emo- One of the first lexicons used in sentiment anal- tion recognition (Polzin and Waibel, 1996; Tato et ysis is the General Inquirer (Stone, 1968). Since al., 2002; Ayadi et al., 2011). then, many methods have been developed to auto- There are also studies that analyzed the visual matically identify opinion words and their polarity cues, such as facial expressions and body move- (Hatzivassiloglou and McKeown, 1997; Turney, ments (Calder et al., 2001; Rosenblum et al., 1996; 2002; Hu and Liu, 2004; Taboada et al., 2011), as Essa and Pentland, 1997). Facial expressions are well as n-gram and more linguistically complex among the most powerful and natural means for phrases (Yang and Cardie, 2012). human beings to communicate their emotions and For data-driven methods, one of the most widely intentions (Tian et al., 2001). Emotions can be used datasets is the MPQA corpus (Wiebe et al., also expressed unconsciously, through subtle move- 2005), which is a collection of news articles manu- ments of facial muscles such as smiling or eyebrow ally annotated for opinions. Other datasets are also raising, often measured and described using the available, including two polarity datasets consist- Facial Action Coding System (FACS) (Ekman et ing of movie reviews (Pang and Lee, 2004; Maas et al., 2002). al., 2011), and a collection of newspaper headlines De Silva et. al. (De Silva et al., 1997) and Chen annotated for polarity (Strapparava and Mihalcea, et. al. (Chen et al., 1998) presented one of the 2007). early works that integrate both acoustic and visual While difficult problems such as cross-domain information for emotion recognition. In addition to (Blitzer et al., 2007; Li et al., 2012) or cross- work that considered individual modalities, there language (Mihalcea et al., 2007; Wan, 2009; Meng is also a growing body of work concerned with et al., 2012) portability have been addressed, not multimodal emotion analysis (Silva et al., 1997; much has been done in terms of extending the ap- Sebe et al., 2006; Zhihong et al., 2009; Wollmer et plicability of sentiment analysis to other modalities, al., 2010). 974 Utterance transcription Label En este color, creo que era el color frambuesa. neu In this color, I think it was raspberry Pinta hermosisimo. pos It looks beautiful. Sinceramente, con respecto a lo que pinta y a que son hidratante, si son muy hidratantes. pos Honestly, talking about how they looks and hydrates, yes they are very hydrant. Pero el problema de estos labiales es que cuando uno se los aplica, te dejan un gusto asqueroso en la boca. neg But the problem with those lipsticks is that when you apply them, they leave a very nasty taste Sinceramente, es no es que sea el olor sino que es mas bien el gusto. neg Honestly, is not the smell, it is the taste. Table 1: Sample utterance-level annotations. The labels used are: pos(itive), neg(ative), neu(tral). More recently, two challenges have been or- Among all the videos returned by the YouTube ganized focusing on the recognition of emotions search, we selected only videos that respected the using audio and visual cues (Schuller et al., following guidelines: the speaker should be in front 2011a; Schuller et al., 2011b), which included sub- of the camera; her face should be clearly visible, challenges on audio-only, video-only, and audio- with a minimum amount of face occlusion during video, and drew the participation of many teams the recording; there should not be any background from around the world.
Recommended publications
  • AUTHOR a Longitudinal Study of the Language Development Of
    DOCUMENT RESUME ED 247 706 EC 170 061 AUTHOR Mohay, Heather TITLE A Longitudinal Study of the Language Development of Three Deaf Children of Hearing Parents. PUB DATE , Jun 81 NOTE 26p.,.; Paper presented at a Seminar on Early Intervention or Young Hearing-Impaired Children (Mt. Gravatt., Queensland, Australia, June 15-16,-1981). For the proceedings, see EC 170 057. PUB TYPE Reports' = Research/Technical (143) EDRS PRICE MF01/PCO2 Plus Postage. DESCRIPTORS. *Hearing.Impairments; Infants; *Language Acquisition; Longitudinal Studies; *Manual Communication; Sign Language; Young Children (F ABSTRACT A longitudinal study followed the language acquisition'of three deaf infants. Analysis of videotapes recorded in the child's home during informal play was performed in terms of communicative gestures. Results revealed that Ss used a very limited number of hand configurations, locations for signs, and hand and arm, movements. Analysis of the gestures showed that there was a great deal of similarity in the components used by the three Ss, Ss presented consistent patterns in their gestures. Although the single gesture utterance was the most common communicative form, all Ss did combine gestures to form two-gesture utterances. Speech was infrequently used, and all Ss had larger gestural than spoken lexitons at 30 months. It was possible for elements in two gesture utterances and word/gesture combinations to be produced simultaneously. The content of the utterances indicated that they were able to express the same semantic relationships as hearing children at a similar stage of language development. (CL) 9 *********************************************************************** Reproductions supplied by EDRS are the best that can be made from the original document.
    [Show full text]
  • Two-Word Utterances Chomsky's Influence
    Two-Word Utterances When does language begin? In the middle 1960s, under the influence of Chomsky’s vision of linguistics, the first child language researchers assumed that language begins when words (or morphemes) are combined. (The reading by Halliday has some illustrative citations concerning this narrow focus on “structure.”) So our story begins with what is colloquially known as the “two-word stage.” The transition to 2-word utterances has been called “perhaps, the single most disputed issue in the study of language development” (Bloom, 1998). A few descriptive points: Typically children start to combine words when they are between 18 and 24 months of age. Around 30 months their utterances become more complex, as they add additional words and also affixes and other grammatical morphemes. These first word-combinations show a number of characteristics. First, they are systematically simpler than adult speech. For instance, function words are generally not used. Notice that the omission of inflections, such as -s, -ing, -ed, shows that the child is being systematic rather than copying. If they were simply imitating what they heard, there is no particular reason why these grammatical elements would be omitted. Conjunctions (and), articles (the, a), and prepositions (with) are omitted too. But is this because they require extra processing, which the child is not yet capable of? Or do they as yet convey nothing to the child—can she find no use for them? Second, as utterances become more complex and inflections are added, we find the famous “over-regularization”—which again shows, of course, that children are systematic, not simply copying what they here.
    [Show full text]
  • The Utterance, and Otherbasic Units for Second Language Discourse Analysis
    The Utterance, and OtherBasic Units for Second Language Discourse Analysis GRAHAM CROOKES University of Hawaii-Manoa Theselection ofa base unitisan importantdecision in theprocess ofdiscourse analysis. A number of different units form the bases of discourse analysis systems designed for dealing with structural characteristics ofsecondlanguage discourse. Thispaperreviews the moreprominentofsuch unitsand provides arguments infavourofthe selection ofone inparticular-the utterance. INTRODUCTION Discourse analysis has been an accepted part of the methodology of second language (SL) research for some time (see, for example, Chaudron 1988: 40-5; Hatch 1978; Hatch and Long 1980; Larsen-Freeman 1980), and its procedures are becoming increasingly familiar and even to some extent standardized. As is well-known, an important preliminary stage in the discourse analysis of speech is the identification of units relevant to the investigation within the body of text to be analysed, or the complete separation of the corpus into those basic units. In carrying out this stage of the analysis, the varying needs of second language discourse analysis have caused investigators to make use ofall of the traditional grammatical units of analysis (morpheme, word, clause, etc.), as well as other structural or interactional features of the discourse (for example, turns), in addition to a variety of other units defined in terms of their functions (for example, moves). These units have formed the bases of a number of different analytic systems, which canmainly beclassified as either structural orfunctional (Chaudron 1988), developed to address differing research objectives. In carrying out structural discourse analyses of oral text, researchers have been confronted with the fact that most grammars are based on a unit that is not defined for speech, but is based on the written mode of language-that is, the sentence.
    [Show full text]
  • Stochastic Language Generation in Dialogue Using Factored Language Models
    Stochastic Language Generation in Dialogue Using Factored Language Models Franc¸ois Mairesse∗ University of Cambridge ∗∗ Steve Young University of Cambridge Most previous work on trainable language generation has focused on two paradigms: (a) using a statistical model to rank a set of pre-generated utterances, or (b) using statistics to determine the generation decisions of an existing generator. Both approaches rely on the existence of a hand- crafted generation component, which is likely to limit their scalability to new domains. The first contribution of this article is to present BAGEL, a fully data-driven generation method that treats the language generation task as a search for the most likely sequence of semantic concepts and realization phrases, according to Factored Language Models (FLMs). As domain utterances are not readily available for most natural language generation tasks, a large creative effort is required to produce the data necessary to represent human linguistic variation for nontrivial domains. This article is based on the assumption that learning to produce paraphrases can be facilitated by collecting data from a large sample of untrained annotators using crowdsourcing—rather than a few domain experts—by relying on a coarse meaning representation. A second contribution of this article is to use crowdsourced data to show how dialogue naturalness can be improved by learning to vary the output utterances generated for a given semantic input. Two data- driven methods for generating paraphrases in dialogue are presented: (a) by sampling from the n-best list of realizations produced by BAGEL’s FLM reranker; and (b) by learning a structured perceptron predicting whether candidate realizations are valid paraphrases.
    [Show full text]
  • Annotation of Utterances for Conversational Nonverbal Behaviors
    Annotation of Utterances for Conversational Nonverbal Behaviors Allison Funkhouser CMU-RI-TR-16-25 Submitted in partial fulfillment of the requirements for the degree of Masters of Science in Robotics May 2016 Masters Committee: Reid Simmons, Chair Illah Nourbakhsh Heather Knight Abstract Nonverbal behaviors play an important role in communication for both humans and social robots. However, hiring trained roboticists and animators to individually animate every possible piece of dialogue is time consuming and does not scale well. This has motivated previous researchers to develop automated systems for inserting appropriate nonverbal behaviors into utterances based only on the text of the dialogue. Yet this automated strategy also has drawbacks, because there is basic semantic information that humans can easily identify that is not yet accurately captured by a purely automated system. Identifying the dominant emotion of a sentence, locating words that should be emphasized by beat gestures, and inferring the next speaker in a turn-taking scenario are all examples of data that would be useful when animating an utterance but which are difficult to determine automatically. This work proposes a middle ground between hand-tuned animation and a purely text-based system. Instead, untrained human workers label relevant semantic information for an utterance. These labeled sentences are then used by an automated system to produce fully animated dialogue. In this way, the relevant human-identifiable context of a scenario is preserved without requiring workers to have deep expertise of the intricacies of nonverbal behavior. Because the semantic information is independent of the robotic platform, workers are also not required to have access to a simulation or physical robot.
    [Show full text]
  • Utterance Structure in Initial L2 Acquisition
    Utterance structure in initial L2 acquisition Jacopo Saturno language Eurosla Studies 2 science press EuroSLA Studies Editor: Gabriele Pallotti Associate editors: Amanda Edmonds, Université de Montpellier; Ineke Vedder, University of Amsterdam In this series: 1. Pérez Vidal, Carmen, Sonia López­Serrano, Jennifer Ament & Dakota J. Thomas­Wilhelm (eds.). Learning context effects: Study abroad, formal instruction and international immersion classrooms 2. Saturno, Jacopo. Utterance structure in initial L2 acquisition. ISSN: 2626­2665 Utterance structure in initial L2 acquisition Jacopo Saturno language science press Saturno, Jacopo. 2020. Utterance structure in initial L2 acquisition (Eurosla Studies 2). Berlin: Language Science Press. This title can be downloaded at: http://langsci-press.org/catalog/book/265 © 2020, Jacopo Saturno Published under the Creative Commons Attribution 4.0 Licence (CC BY 4.0): http://creativecommons.org/licenses/by/4.0/ ISBN: 978-3-96110-261-7 (Digital) 978-3-96110-262-4 (Hardcover) ISSN: 2626-2665 DOI: 10.5281/zenodo.3889998 Source code available from www.github.com/langsci/265 Collaborative reading: paperhive.org/documents/remote?type=langsci&id=265 Cover and concept of design: Ulrike Harbort Typesetting: Sebastian Nordhoff Proofreading: Amir Ghorbanpour, Andreas Hölzl, Brett Reynolds, Jeroen van de Weijer, Lachlan Mackenzie, Sebastian Nordhoff, Tom Bossuyt Fonts: Libertinus, Arimo, DejaVu Sans Mono Typesetting software:Ǝ X LATEX Language Science Press Xhain Grünberger Str. 16 10243 Berlin, Germany langsci-press.org Storage and cataloguing done by FU Berlin Contents Acknowledgements v 1 Introduction: Rationale and research questions 1 1.1 Word order vs. inflectional morphology .............. 3 1.2 Task effects: structured tests vs. semi-spontaneous production .
    [Show full text]
  • An Analysis of Code Switching in the Novel 9 Summers 10 Autumns
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Vivid Journal of Language and Literature AN ANALYSIS OF CODE SWITCHING IN THE NOVEL 9 SUMMERS 10 AUTUMNS Wenny Yuliani [email protected], 0810733112 English Department, Faculty of Humanities, Andalas University ABSTRAK Kajian ini membahas tipe dan fungsi alih kode dalam percakapan-percakapan pada novel 9 Summers 10 Autumns karya Iwan Setyawan. Metode yang digunakan untuk mengumpulkan data adalah simak bebas libat cakap (Sudaryanto 1993), sedangkan teori yang diterapkan dalam menganalisis data adalah teori dari Gumperz (1982) dan Hoffman (1996). Hasil analisis data dipaparkan dengan menggunakan metode formal dan informal. Dari data yang dianalisis ditemukan tiga tipe alih kode yang muncul dalam ujaran, yaitu 14 inter-sentential switching, 4 intra-sentential switching dan 2 emblematic switching. Selain itu ditemukan empat fungsi alih kode, yakni 7 message qualification, 5 adressee specification, 5 personalization, dan 3 interjection. Dari enam fungsi alih kode yang dikemukakan Gumperz, fungsi alih kode quotation dan reiteration tidak muncul dalam data. Hal ini disebabkan karena karakter tidak pernah mengutip ujaran orang lain ketika dia berujar dengan penutur lain dan juga tidak mengulang ujaran dari satu bahasa ke bahasa lainya. Kata kunci: alih kode, tipe alih kode, fungsi alih kode ABSTRACT This study is aimed to describe the types and the functions of code switching which are uttered by the characters in the novel 9 Summers 10 Autumns. In collecting the data, non participatory observational method (Sudaryanto 1993) was applied. In analyzing the data, Hoffman’s theory (1996) is referred to analyze the types of code switching and Gumperz’s theory (1982) is referred to analyze the functions of the code switching.
    [Show full text]
  • The Semantics of Onomatopoeic Speech Act Verbs
    The Semantics of Onomatopoeic Speech Act Verbs I-Ni Tsai Chu-Ren Huang Institute of Linguistics (Preparatory Office), Institute of Linguistics (Preparatory Office), Academia Sinica Academia Sinica No. 128 Academic Sinica Road, Sec 2. No. 128 Academic Sinica Road, Sec 2. Nangkang, Taipei 115, Taiwan Nangkang, Taipei 115, Taiwan [email protected] [email protected] Abstract This paper attempts to explore the semantics of onomatopoeic speech act verbs. Language abounds with small bits of utterances to show speaker's emotions, to maintain the flow of speech and to do some daily exchange routines. These tiny vocalizations have been regarded as vocal gestures and largely studied under the framework of 'interjection'. In this paper, the emphasis is placed on the perlocutionary force the vocal tokens contain. We describe their conventionalized lexical meaning and term them as onomatopoeic speech act verb. An onomatopoeic speech act verb refers to a syntactically independent monomorphemic utterance which performs illocutionary or perlocutionary forces. It is normally directed at the listener, which making the recipient to do something or to solicit recipient's response or reaction. They are onomatopoeic because most of them are imitation of the sounds produced by doing some actions. 1 Introduction Based on the Speech Act Theory (Austin 1992; Reiss 1985; Searle 1969, 1975, 1979), speech act is what speaker performs when producing the utterance. Researchers suggest that speech act is not only an utterance act but contains the perlocutionary force. If the speaker has some particular intention when making the utterance such as committing to doing something or expressing attitude or emotions, the speech act is said to contain illocutionary force.
    [Show full text]
  • Chapter 9. Utterances and Their Meanings
    CHAPTER 9 UTTERANCES AND THEIR MEANINGS Meanings are constructed situationally by the participants in interaction, as they construe intent in each other’s uttered words. A well-known story (said to be a favorite of both Vygotsky and Bakhtin) tells of a group of sailors having a nuanced exchange by repeating the same expletive to each other, but with a different intonation and timing at each turn. This polysemousness of words is equally to be found in an office memo announcing a change in reporting procedures that leaves the recipients wondering what the real meaning is—from enacting a corporate shake-up, to disciplining a co-worker, to a power-grab by a manager, to simply creating an efficiency. Much water-cooler time may be devoted to examining the nuances of expression or sharing other contexting information until a stable social meaning is agreed on, which will then guide the behavior of all concerned. To put it explicitly, meaning is not a property of language in itself, and is not immanent in language. Meaning is what people construe using the prosthesis of language, interpreted within specific contexts of use. To understand meaning, we need to take utterance and people’s construal of utterance as our fundamental units of analysis. VOLOSINOV AND HIS CIRCLE’S PROPOSAL FOR AN UTTERANCE-BASED LINGUISTICS Volosinov in Marxism and the Philosophy of Language (1929/1973), foreshadowed by comments in his earlier work on Freud (1927/1987), argued that linguistics should be grounded in utterance, rather than in the formal structure of language. Utterance was the natural unit of speech and communication, with each utterance taking shape within a recognizable form (that is, a speech genre), directed to a specific audience (what Bakhtin, 1984a, 1986, was to call addressivity), and in response to prior utterances.
    [Show full text]
  • Utterance Production
    Utterance Production LCSD56 Psycholinguistics Production It's rather nice in the sense that where the mews is on the way up to Kilburn it's a cheaper area it's not a select shopping centre by any means and there are lots of council houses and flats and I think it's fantastic because they're very nice looking flats and everything has been fairly well designed and you can go up there and shop reasonably 2 Production – actual delivery Its just + rather nice in the sense that where ++ the mews is erm ++ on ++ on the way up to Kilburn + you've got th- ++ well its its more of a er a cheaper + its not a select shopping centre by any means and there're lots of + council houses and flats and ++ erm + I mean I + think its del- its fantastic because you can go up there and they're very + nice looking flats and everything its its been fairly well designed + 'nd you can go up there and + and shop reasonably. What are the main differences in pausing between actual and ideal delivery? Is there any pattern to the repetitions? 3 Production – actual delivery Its just + rather nice in the sense that where ++ the mews is erm ++ on ++ on the way up to Kilburn + you've got th- ++ well its its more of a er a cheaper + its not a select shopping centre by any means and there're lots of + council houses and flats and ++ erm + I mean I + think its del- its fantastic because you can go up there and they're very + nice looking flats and everything its its been fairly well designed + 'nd you can go up there and + and shop reasonably.
    [Show full text]
  • The Neat Summary of Linguistics
    The Neat Summary of Linguistics Table of Contents Page I Language in perspective 3 1 Introduction 3 2 On the origins of language 4 3 Characterising language 4 4 Structural notions in linguistics 4 4.1 Talking about language and linguistic data 6 5 The grammatical core 6 6 Linguistic levels 6 7 Areas of linguistics 7 II The levels of linguistics 8 1 Phonetics and phonology 8 1.1 Syllable structure 10 1.2 American phonetic transcription 10 1.3 Alphabets and sound systems 12 2 Morphology 13 3 Lexicology 13 4 Syntax 14 4.1 Phrase structure grammar 15 4.2 Deep and surface structure 15 4.3 Transformations 16 4.4 The standard theory 16 5 Semantics 17 6 Pragmatics 18 III Areas and applications 20 1 Sociolinguistics 20 2 Variety studies 20 3 Corpus linguistics 21 4 Language and gender 21 Raymond Hickey The Neat Summary of Linguistics Page 2 of 40 5 Language acquisition 22 6 Language and the brain 23 7 Contrastive linguistics 23 8 Anthropological linguistics 24 IV Language change 25 1 Linguistic schools and language change 26 2 Language contact and language change 26 3 Language typology 27 V Linguistic theory 28 VI Review of linguistics 28 1 Basic distinctions and definitions 28 2 Linguistic levels 29 3 Areas of linguistics 31 VII A brief chronology of English 33 1 External history 33 1.1 The Germanic languages 33 1.2 The settlement of Britain 34 1.3 Chronological summary 36 2 Internal history 37 2.1 Periods in the development of English 37 2.2 Old English 37 2.3 Middle English 38 2.4 Early Modern English 40 Raymond Hickey The Neat Summary of Linguistics Page 3 of 40 I Language in perspective 1 Introduction The goal of linguistics is to provide valid analyses of language structure.
    [Show full text]
  • Redalyc.MEAN LENGTH of UTTERANCE-WORDS IN
    Revista CEFAC ISSN: 1516-1846 [email protected] Instituto Cefac Brasil Santos, Maria Emília; Lynce, Sofia; Carvalho, Sara; Cacela, Mariana; Mineiro, Ana MEAN LENGTH OF UTTERANCE-WORDS IN CHILDREN WITH TYPICAL LANGUAGE DEVELOPMENT AGED 4 TO 5 YEARS Revista CEFAC, vol. 17, núm. 4, julio-agosto, 2015, pp. 1143-1151 Instituto Cefac São Paulo, Brasil Available in: http://www.redalyc.org/articulo.oa?id=169342207012 How to cite Complete issue Scientific Information System More information about this article Network of Scientific Journals from Latin America, the Caribbean, Spain and Portugal Journal's homepage in redalyc.org Non-profit academic project, developed under the open access initiative 1143 MEAN LENGTH OF UTTERANCE-WORDS IN CHILDREN WITH TYPICAL LANGUAGE DEVELOPMENT AGED 4 TO 5 YEARS Extensão média do enunciado-palavras em crianças de 4 e 5 anos com desenvolvimento típico da linguagem Maria Emília Santos (1), Sofia Lynce (1), Sara Carvalho (1), Mariana Cacela (1), Ana Mineiro (1) ABSTRACT Purpose: to analyze the Mean Length of Utterance-words (MLU-w) in children aged 4;00-5;05 years. Methods: ninety two Portuguese children with normal development were observed: 49 girls and 43 boys, divided in age range groups of six months. A sample of 100 utterances produced in spontaneous discourse was collected from each child. The utterances were transcribed and analyzed. Results: MLU-w was shown to vary between 4,5 to 5 words, progressing with age. This progression had been previously observed in US English and in Brazilian Portuguese speaking children, although in European Portuguese the number of words is overall a little higher.
    [Show full text]