Emotion Algebra Reveals That Sequences of Facial Expressions Have Meaning Beyond Their Discrete Constituents

Total Page:16

File Type:pdf, Size:1020Kb

Emotion Algebra Reveals That Sequences of Facial Expressions Have Meaning Beyond Their Discrete Constituents Emotion Algebra reveals that sequences of facial expressions have meaning beyond their discrete constituents Carmel Sofer1, Dan Vilenchik2, Ron Dotsch3, Galia Avidan1 1. Ben Gurion University of the Negev, Department of Psychology 2. Ben Gurion University of the Negev, Department of Communication Systems Engineering 3. Utrecht University, Department of Psychology 1 Abstract Studies of emotional facial expressions reveal agreement among observes about the meaning of six to fifteen basic static expressions. Other studies focused on the temporal evolvement, within single facial expressions. Here, we argue that people infer a larger set of emotion states than previously assumed, by taking into account sequences of different facial expressions, rather than single facial expressions. Perceivers interpreted sequences of two images, derived from eight facial expressions. We employed vector representation of emotion states, adopted from the field of Natural Language Processing and performed algebraic vector computations. We found that the interpretation participants ascribed to the sequences of facial expressions could be expressed as a weighted average of the single expressions comprising them, resulting in 32 new emotion states. Additionally, observes’ agreement about the interpretations, was expressed as a multiplication of respective expressions. Our study sheds light on the significance of facial expression sequences to perception. It offers an alternative account as to how a variety of emotion states are being perceived and a computational method to predict these states. Keywords: word embedding, emotions, facial expressions, vector representation 2 Studies of emotional facial expressions reveal agreement among observers about the meaning of basic expressions, ranging between six to fifteen emotional states (Ekman, 1994, 2016; Ekman & Cordaro, 2011; Keltner & Cordaro, 2015; Russell, 1994), creating a vocabulary conveyed during face to face interactions. Other studies which used a different approach found 21 expression categories which combine muscle movements of two expressions (e.g., a single image of sadly fearful expression combined muscle movements of sad and fear expressions; Du, Tao, & Martinez, 2014) . However, if the vocabulary of the emotion-states is indeed limited to that small number of discrete facial expressions, it would have limited people’s ability to express and perceive finer emotions and emotional states. Thus, it is conceivable that people can perceive a much wider collection of emotion states, using a combination of basic, discrete facial expressions. Such mechanisms, which expand a very small number of basic components to a much wider number of meanings, exist in other human communications modalities, such as spoken language and music. For example, in the English language – a combination of 26 letters, forms a rich language of more than 200,000 words (Merriam- Webster, 2002) and an unlimited number of sentences; More so in western music; a combination of seven notes creates endless musical compositions. Here, we argue that people infer a larger set of emotion states from facial expressions than previously assumed, by taking into account sequences of facial expressions, rather than single facial expressions. Some studies already revealed that an expression presented to participants affects their interpretation of subsequent facial expressions (Russell, 1991; Russell & Fehr, 1987; Thayer, 1980). For example, Russell (1991) found that a contempt facial 3 expression, followed by a sadness expression was interpreted as disgust. However, when a disgust expression was presented first, the contempt facial expression was perceived as a sign of sadness. Other studies which focused on the temporal evolvement within single facial expressions (lasting one to two seconds), have found that early in their evolvement, facial expressions allow for crude differentiation of expression cues, while later transmissions consist of more complex cues that support finer categorization (Jack, Garrod, & Schyns, 2014). Critically, these previous studies were constrained by focusing on effects of single facial expressions or on the effect of specific facial expressions on the interpretation of subsequent expressions, while a significant social interaction usually includes various facial expressions. As such, they might have overlooked a more general system of rules governing the interpretation of sequences of facial expressions as a whole, whereby sequences of basic facial expressions are mapped onto specific integrated interpretations of emotion states. For example, would a person displaying a surprise, followed by a smile, followed by sadness facial expressions, be perceived as surprised? happy? sad? Or alternatively the integrated interpretation of the sequence, would be helplessness, a breakdown or maybe disappointment. Several processes might affect one’s interpretation of a sequence of expressions. One possibility is that during a face to face interaction, observers accumulate the emotional information conveyed by facial expressions. Emotional information is integrated (summed), taking into account the different facial expressions and their duration. Another alternative is that weights are assigned to discrete events (“snapshots”), based on their location in the sequence, irrespective of their duration 4 (Fredrickson & Kahneman, 1993). The resulted impression would then be the weighted average of the judgments assigned to these snapshots. Studies that focused on scene perception support the latter alternative (Fredrickson & Kahneman, 1993; McDuff, El Kaliouby, Cohn, & Picard, 2015; Noë et al., 2017; Varey & Kahneman, 1992). The goal of this study is to model, using a data-driven approach, how the order of serially presented facial expressions, map onto a perceiver’s interpretation of the sequence. We adopted language modeling and feature learning methods from the field of Natural Language Processing (NLP), which rely on computational methodologies and machine learning techniques. There are various approaches used in this domain, for example, the word2vec family of methods (Mikolov, Chen, Corrado, & Dean, 2013)and the GloVe - (Global Vectors for Word Representation; Pennington, Socher, & Manning, 2014). Generally, the application of these methods, termed word embedding, maps words from a text corpus to numerical vectors. Each word in the corpus is represented as a vector in a multi-dimensional word space, enabling the usage of vector arithmetic. For example, one can perform a simple and useful "emotion algebra", by averaging observers’ interpretation of “happy” facial expression (see Figure 1). The expression may be perceived as a sign of happiness by one participant or satisfaction by another, hence represented by the vectors of the words happy or satisfaction respectively. However, averaging the judgment vectors of all participants, may reveal that on average, the “happy” expression is closest to the vector of the word excitement. By using word embedding methods, here we integrated a mathematical framework with participants’ empirical data, arriving at conclusions, which would have been hard to achieve with traditional statistical approaches. 5 We examined 3 hypotheses: (i) The meaning of sequences of emotional facial expressions is manifested by the weighted average of its discrete constituents, rather than their integration (summation). (ii) The size of the extended vocabulary of emotion expression, based on sequences of emotions, is larger than the number of their discrete constituents, and smaller than the upper bound of all possible theoretical combinations of facial expressions. (iii) Frequent sequences of facial expressions that are more likely to appear in real life (Seger, 1994) elicit greater agreement among observers compared to other, less frequent sequences. We predict that this agreement, reflected in participants’ judgment variability, can be modeled by vector arithmetics. Figure 1: A 2D (D1, D2) illustration of the averaging process of participants’ interpretation of the “happy” expression, perceived by participants as signs of happiness or satisfaction. By applying vector arithmetics, we find that, on average, participants judge the “happy” expression as a sign of excitement (dashed black line). To examine these 3 hypotheses, participants were presented with 64 sequences of two facial expressions. They were required to describe, using a single word in free text format, the “state of mind” of the person in the picture, based on all expressions in the sequence. 6 Methods Participants Forty-seven participants1 fluent in English (19 female participants) aged 20 to 70 (M= 36.65) took part in an on-line study, conducted from their homes at their own pace using Amazon’s Mechanical-Turk platform. Before the study, participants were requested to repeat in writing the instructions given to them, using their own words. Their descriptions, reviewed independently by two researchers served as an indication for their language fluency and comprehension of the task. Seven participants were excluded from the study because their language was unreasonably not fluent. Participants received a payment of 1 USD, for their participation, which lasted 27 minutes on average. Stimuli Stimuli included eight facial expressions consisted of seven most consensual expressions: anger, contempt, disgust, happiness, fear, sadness and surprise (Ekman & Friesen, 1971) complemented by a neutral expression. Images were created by digitally averaging,
Recommended publications
  • The Development of Emotion Recognition from Facial Expressions and Non‐Linguistic Vocalizations During Childhood
    1 British Journal of Developmental Psychology (2014) © 2014 The British Psychological Society www.wileyonlinelibrary.com The development of emotion recognition from facial expressions and non-linguistic vocalizations during childhood Georgia Chronaki1,2*, Julie A. Hadwin2, Matthew Garner3, Pierre Maurage4 and Edmund J. S. Sonuga-Barke2,5 1Section of Clinical and Cognitive Neuroscience, School of Psychological Sciences, University of Manchester, UK 2Developmental Brain-Behaviour Laboratory, Psychology, University of Southampton, UK 3Psychology, Faculty of Social & Human Sciences & Clinical and Experimental Sciences, Faculty of Medicine, University of Southampton, UK 4Laboratory for Experimental Psychopathology, Psychological Sciences Research Institute, Catholic University of Louvain, Louvain-la-Neuve, Belgium 5Department of Experimental Clinical and Health Psychology, Ghent University, Belgium Sensitivity to facial and vocal emotion is fundamental to children’s social competence. Previous research has focused on children’s facial emotion recognition, and few studies have investigated non-linguistic vocal emotion processing in childhood. We compared facial and vocal emotion recognition and processing biases in 4- to 11-year-olds and adults. Eighty-eight 4- to 11-year-olds and 21 adults participated. Participants viewed/listened to faces and voices (angry, happy, and sad) at three intensity levels (50%, 75%, and 100%). Non-linguistic tones were used. For each modality, participants completed an emotion identification task. Accuracy and bias for each emotion and modality were compared across 4- to 5-, 6- to 9- and 10- to 11-year-olds and adults. The results showed that children’s emotion recognition improved with age; preschoolers were less accurate than other groups. Facial emotion recognition reached adult levels by 11 years, whereas vocal emotion recognition continued to develop in late childhood.
    [Show full text]
  • Recognising Spontaneous Facial Micro-Expressions
    Recognising Spontaneous Facial Micro-expressions Tomas Pfister, Xiaobai Li, Guoying Zhao and Matti Pietik¨ainen Machine Vision Group, Department of Computer Science and Engineering, University of Oulu PO Box 4500, 90014 Oulu, Finland {tpfister,lxiaobai,gyzhao,mkp}@ee.oulu.fi Abstract Facial micro-expressions are rapid involuntary facial ex- pressions which reveal suppressed affect. To the best knowl- edge of the authors, there is no previous work that success- fully recognises spontaneous facial micro-expressions. In this paper we show how a temporal interpolation model together with the first comprehensive spontaneous micro- expression corpus enable us to accurately recognise these very short expressions. We designed an induced emotion Figure 1. An example of a facial micro-expression (top-left) be- suppression experiment to collect the new corpus using a ing interpolated through graph embedding (top-right); the result high-speed camera. The system is the first to recognise from which spatiotemporal local texture descriptors are extracted spontaneous facial micro-expressions and achieves very (bottom-right), enabling recognition with multiple kernel learning. promising results that compare favourably with the human micro-expression detection accuracy. proposed a suitable price. Since the human recognition accuracy is so low, an alternative method for recognising 1. Introduction micro-expressions would be very valuable. Humans are good at recognising full facial expressions The major challenges in recognising micro-expressions which present a rich source of affective information [6]. involve their very short duration and involuntariness. The However, psychological studies [4, 8] have shown that af- short duration means only a very limited number of frames fect also manifests itself as micro-expressions.
    [Show full text]
  • Visual Perception of Facial Emotional Expressions During Saccades
    behavioral sciences Article Visual Perception of Facial Emotional Expressions during Saccades Vladimir A. Barabanschikov and Ivan Y. Zherdev * Institute of Experimental Psychology, Moscow State University of Psychology and Education, 29 Sretenka street, Moscow 127051, Russia; [email protected] * Correspondence: [email protected]; Tel.: +7-999-829-52-79 Received: 28 October 2019; Accepted: 23 November 2019; Published: 27 November 2019 Abstract: The regularities of visual perception of both complex and ecologically valid objects during extremely short photo expositions are studied. Images of a person experiencing basic emotions were displayed for as low as 14 ms amidst a saccade spanning 10 degrees of visual angle. The observers had a main task to recognize the emotion depicted, and a secondary task to point at the perceived location of the photo on the screen. It is shown that probability of correct recognition of emotion is above chance (0.62), and that it depends on its type. False localizations of stimuli and their compression in the direction of the saccade were also observed. According to the acquired data, complex environmentally valid objects are perceived differently during saccades in comparison to isolated dots, lines or gratings. The rhythmic structure of oculomotor activity (fixation–saccade–fixation) does not violate the continuity of the visual processing. The perceptual genesis of facial expressions does not take place only during gaze fixation, but also during peak speed of rapid eye movements both at the center and in closest proximity of the visual acuity area. Keywords: transsaccadic processing; facial expression; emotional perception; saccadic suppression; gaze contingency; visual recognition 1.
    [Show full text]
  • Facial Expression and Emotion Paul Ekman
    1992 Award Addresses Facial Expression and Emotion Paul Ekman Cross-cultural research on facial expression and the de- We found evidence of universality in spontaneous velopments of methods to measure facial expression are expressions and in expressions that were deliberately briefly summarized. What has been learned about emo- posed. We postulated display rules—culture-specific pre- tion from this work on the face is then elucidated. Four scriptions about who can show which emotions, to whom, questions about facial expression and emotion are dis- and when—to explain how cultural differences may con- cussed: What information does an expression typically ceal universal in expression, and in an experiment we convey? Can there be emotion without facial expression? showed how that could occur. Can there be a facial expression of emotion without emo- In the last five years, there have been a few challenges tion? How do individuals differ in their facial expressions to the evidence of universals, particularly from anthro- of emotion? pologists (see review by Lutz & White, 1986). There is. however, no quantitative data to support the claim that expressions are culture specific. The accounts are more In 1965 when I began to study facial expression,1 few- anecdotal, without control for the possibility of observer thought there was much to be learned. Goldstein {1981} bias and without evidence of interobserver reliability. pointed out that a number of famous psychologists—F. There have been recent challenges also from psychologists and G. Allport, Brunswik, Hull, Lindzey, Maslow, Os- (J. A. Russell, personal communication, June 1992) who good, Titchner—did only one facial study, which was not study how words are used to judge photographs of facial what earned them their reputations.
    [Show full text]
  • Facial Micro-Expression Analysis Jingting Li
    Facial Micro-Expression Analysis Jingting Li To cite this version: Jingting Li. Facial Micro-Expression Analysis. Computer Vision and Pattern Recognition [cs.CV]. CentraleSupélec, 2019. English. NNT : 2019CSUP0007. tel-02877766 HAL Id: tel-02877766 https://tel.archives-ouvertes.fr/tel-02877766 Submitted on 22 Jun 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. THESE DE DOCTORAT DE CENTRALESUPELEC COMUE UNIVERSITE BRETAGNE LOIRE ECOLE DOCTORALE N° 601 Mathématiques et Sciences et Technologies de l'Information et de la Communication Spécialité : Signal, Image, Vision Par « Jingting LI » « Facial Micro-Expression Analysis » Thèse présentée et soutenue à « Rennes », le « 02/12/2019 » Unité de recherche : IETR Thèse N° : 2019CSUP0007 Rapporteurs avant soutenance : Composition du Jury : Olivier ALATA Professeur, Université Jean Monnet, Président / Rapporteur Saint-Etienne Olivier ALATA Professeur, Fan YANG-SONG Professeur, Université de Bourgogne, Université Jean Monnet, Dijon Saint-Etienne Rapporteuse Fan YANG-SONG Professeur, Université de Bourgogne,
    [Show full text]
  • Gestures: Your Body Speaks
    GESTURES: YOUR BODY SPEAKS How to Become Skilled WHERE LEADERS in Nonverbal Communication ARE MADE GESTURES: YOUR BODY SPEAKS TOASTMASTERS INTERNATIONAL P.O. Box 9052 • Mission Viejo, CA 92690 USA Phone: 949-858-8255 • Fax: 949-858-1207 www.toastmasters.org/members © 2011 Toastmasters International. All rights reserved. Toastmasters International, the Toastmasters International logo, and all other Toastmasters International trademarks and copyrights are the sole property of Toastmasters International and may be used only with permission. WHERE LEADERS Rev. 6/2011 Item 201 ARE MADE CONTENTS Gestures: Your Body Speaks............................................................................. 3 Actions Speak Louder Than Words...................................................................... 3 The Principle of Empathy ............................................................................ 4 Why Physical Action Helps........................................................................... 4 Five Ways to Make Your Body Speak Effectively........................................................ 5 Your Speaking Posture .................................................................................. 7 Gestures ................................................................................................. 8 Why Gestures? ...................................................................................... 8 Types of Gestures.................................................................................... 9 How to Gesture Effectively..........................................................................
    [Show full text]
  • The Role of Facial Expression in Intra-Individual and Inter-Individual Emotion Regulation
    From: AAAI Technical Report FS-01-02. Compilation copyright © 2001, AAAI (www.aaai.org). All rights reserved. The Role of Facial Expression in Intra-individual and Inter-individual Emotion Regulation Susanne Kaiser and Thomas Wehrle University of Geneva, Faculty of Psychology and Education 40 Bd. du Pont d'Arve, CH-1205 Geneva, Switzerland [email protected] [email protected] Abstract Ellsworth (1991) has argued, facial expressions have been This article describes the role and functions of facial discrete emotion theorists’ major ‘evidence’ for holistic expressions in human-human and human-computer emotion programs that could not be broken down into interactions from a psychological point of view. We smaller units. However, even though universal prototypical introduce our theoretical framework, which is based on patterns have been found for the emotions of happiness, componential appraisal theory and describe the sadness, surprise, disgust, anger, and fear, these findings methodological and theoretical challenges to studying the have not enabled researchers to interpret facial expressions role of facial expression in intra-individual and inter- as unambiguous indicators of emotions in spontaneous individual emotion regulation. This is illustrated by some interactions. results of empirical studies. Finally we discuss the potential significance of componential appraisal theory for There are a variety of problems. First, the mechanisms diagnosing and treating affect disturbances and sketch linking facial expressions to emotions are not known. possible implications for emotion synthesis and affective Second, the task of analyzing the ongoing facial behavior user modeling. in dynamically changing emotional episodes is obviously more complex than linking a static emotional expression to a verbal label.
    [Show full text]
  • Facial Expression Learning Tion Expressions As Adults
    Comp. by: DMuthuKumar Stage: Galleys Chapter No.: 1925 Title Name: ESL Page Number: 0 Date:4/6/11 Time:11:25:27 1 F the same universal and distinguishable prototypic emo- 39 2 Facial Expression Learning tion expressions as adults. These prototypic expressions 40 are produced by affect programs. Affect programs are 41 1 2 Au1 3 DANIEL S. MESSINGER ,EILEEN OBERWELLAND ,WHITNEY neurophysiological mechanisms that link subjectively felt 42 3 3 4 MATTSON ,NAOMI EKAS emotions to facial expressions in an invariant fashion 43 1 5 University of Miami, Coral Gables, FL, USA across the lifespan. Changes in facial expressions are 44 2 6 Leiden, The Netherlands thought to be due to maturation and the influence of 45 3 7 Coral Gables, FL, USA societal display rules on underlying expressions. 46 Cognitive theories of facial expression suggest that 47 newborns begin life with three primary emotion expres- 48 8 Synonyms sions: distress, pleasure, and interest. As cognitive func- 49 9 Emotional Expression Development tions grow in complexity across development, these facial 50 expressions become more differentiated in their presenta- 51 10 Definition tion and more tightly linked to specific contexts. The 52 11 Facial expression learning. Facial expressions are produced cognitive concepts that are necessary to develop the basic 53 12 as the muscles of the face contract, creating facial config- emotions of surprise, interest, joy, anger, sadness, fear, and 54 13 urations that serve communicative and emotional func- disgust develop over the first 6–8 months of life and 55 14 tions. Facial expression learning involves changes in the consist of perceptual and representational abilities.
    [Show full text]
  • A Review on Facial Micro-Expressions Analysis: Datasets, Features and Metrics
    PREPRINT SUBMITTED TO IEEE JOURNAL. 1 A Review on Facial Micro-Expressions Analysis: Datasets, Features and Metrics Walied Merghani, Adrian K. Davison, Member, IEEE, Moi Hoon Yap, Member, IEEE (This work has been submitted to the IEEE for possible publication. Copyright may be transferred without notice, after which this version may no longer be accessible). Abstract—Facial micro-expressions are very brief, spontaneous facial expressions that appear on the face of humans when they either deliberately or unconsciously conceal an emotion. Micro-expression has shorter duration than macro-expression, which makes it more challenging for human and machine. Over the past ten years, automatic micro-expressions recognition has attracted increasing attention from researchers in psychology, computer science, security, neuroscience and other related disciplines. The aim of this paper is to provide the insights of automatic micro-expressions analysis and recommendations for future research. There has been a lot of datasets released over the last decade that facilitated the rapid growth in this field. However, comparison across different datasets is difficult due to the inconsistency in experiment protocol, features used and evaluation methods. To address these issues, we review the datasets, features and the performance metrics deployed in the literature. Relevant challenges such as the spatial temporal settings during data collection, emotional classes versus objective classes in data labelling, face regions in data analysis, standardisation of metrics and the requirements for real-world implementation are discussed. We conclude by proposing some promising future directions to advancing micro-expressions research. Index Terms—Facial micro-expressions, micro-expressions recognition, micro-movements detection, feature extraction, deep learning.
    [Show full text]
  • Facial Expression Recognition Synonyms Definition
    Comp. by: SunselvakumarGalleys0000870950 Date:27/11/08 Time:14:34:41 Stage:First Proof File Path://ppdys1108/Womat3/Production/PRODENV/0000000005/0000010929/0000000016/ 0000870950.3D Proof by: QC by: F ambient intelligence, and ▶ human computing) will Facial Expression Recognition need to develop human-centered ▶ user interfaces that respond readily to naturally occurring, multimod- MAJA PANTIC al, human communication [1]. These interfaces will Imperial College London, London, UK; University of need the capacity to perceive and understand inten- Twente, AE Enschede, The Netherlands tions and emotions as communicated by social and affective signals. Motivated by this vision of the future, automated analysis of nonverbal behavior, and espe- Synonyms cially of facial behavior, has attracted increasing atten- tion in computer vision, pattern recognition, and Facial Expression Analysis; Facial Action Coding human-computer interaction [2–5]. To wit, facial ex- pression is one of the most cogent, naturally preeminent means for human beings to communicate emotions, Definition to clarify and stress what is said, to signal comprehen- sion, disagreement, and intentions, in brief, to regulate Facial expression recognition is a process performed by interactions with the environment and other persons in humans or computers, which consists of: the vicinity [6, 7]. Automatic analysis of facial expres- 1. Locating faces in the scene (e.g., in an image; this sions forms, therefore, the essence of numerous next- step is also referred to as face detection), generation-computing tools including ▶ affective 2. Extracting facial features from the detected face computing technologies (proactive and affective user region (e.g., detecting the shape of facial compo- interfaces), learner-adaptive tutoring systems, patient- nents or describing the texture of the skin in a facial profiled personal wellness technologies, etc.
    [Show full text]
  • Do Infants Express Discrete Emotions? Adult Judgments of Facial, Vocal, and Body Actions
    DO INFANTS EXPRESS DISCRETE EMOTIONS? ADULT JUDGMENTS OF FACIAL, VOCAL, AND BODY ACTIONS By: Linda A. Camras, Jean Sullivan, and George Michel Camras, LA, Sullivan, J, & Michel, GF. Adult judgments of infant expressive behavior: Facial, vocal, and body actions. Journal of Nonverbal Behavior. 1993; 17(3):171-186. Made available courtesy of Springer Verlag: http://www.springerlink.com/content/104925/?p=c13af7a4bec84f28b51fa65a0e061626&pi=0 The original publication is available at www.springerlink.com ***Note: Figures may be missing from this format of the document Abstract: Adult judges were presented with videotape segments showing an infant displaying facial configurations hypothesized to express discomfort/pain, anger, or sadness according to differential emotions theory (Izard, Dougherty, & Hembree, 1983). The segments also included the infant's nonfacial behavior and aspects of the situational context. Judges rated the segments using a set of emotion terms or a set of activity terms. Results showed that judges perceived the discomfort/pain and anger segments as involving one or more negative emotions not predicted by differential emotions theory. The sadness segments were perceived as involving relatively little emotion overall. Body activity accompanying the discomfort/pain and anger configurations was judged to be more jerky and active than body activity accompanying the sadness configurations. The sadness segments were accompanied by relatively little body movement overall. The results thus fail to conform to the predictions of differential emotions theory but provide information that may contribute to the development of a theory of infant expressive behavior. Article: Controversy exists as to whether infants experience and express discrete negative emotions. Early views of emotional development (Bridges, 1932; Werner, 1948) posited that infants initially respond similarly to any negative emotion stimulus.
    [Show full text]
  • Facial Expressions of Emotion Influence Interpersonal Trait Inferences
    FACIAL EXPRESSIONS OF EMOTION INFLUENCE INTERPERSONAL TRAIT INFERENCES Brian Knutson ABSTRACT.. Theorists have argued that facial expressions of emotion serve the inter- personal function of allowing one animal to predict another's behavior. Humans may extend these predictions into the indefinite future, as in the case of trait infer- ence. The hypothesis that facial expressions of emotion (e.g., anger, disgust, fear, happiness, and sadness) affect subjects' interpersonal trait inferences (e.g., domi- nance and affiliation) was tested in two experiments. Subjects rated the disposi- tional affiliation and dominance of target faces with either static or apparently mov- ing expressions. They inferred high dominance and affiliation from happy expressions, high dominance and low affiliation from angry and disgusted expres- sions, and low dominance from fearful and sad expressions. The findings suggest that facial expressions of emotion convey not only a targets internal state, but also differentially convey interpersonal information, which could potentially seed trait inference. What do people infer from facial expressions of emotion? Darwin (1872/1962) suggested that facial muscle movements which originally sub- served individual survival problems (e.g., spitting out noxious food, shield- ing the eyes) eventually allowed animals to predict the behavior of their conspecifics. Current emotion theorists agree that emotional facial expres- sions can serve social predictive functions (e.g., Ekman, 1982, Izard, 1972; Plutchik, 1980). For instance, Frank (1988) hypothesizes that a person might signal a desire to cooperate by smiling at another. Although a viewer may predict a target's immediate behavior on the basis of his or her facial expressions, the viewer may also extrapolate to the more distant future, as in the case of inferring personality traits.
    [Show full text]