Emotion Assessment for Affective Computing Based on Physiological Responses
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Neural Networks for Emotion Classification
Neural Networks for Emotion Classification For obtaining master degree In computer science At Leiden Institute of Advanced Computer Science Under supervision of Dr. Michael S. Lew and Dr. Nicu Sebe by Yafei Sun August 2003 1 ACKNOWLEDGMENTS I would like to express my heartfelt gratitude to Dr. Michael S. Lew for his support, invaluable guidance, time, and encouragement. I’d like to express my sincere appreciation to Dr. Nicu Sebe for his countless ideas and advice, his encouragement and sharing his valuable knowledge with me. Thanks to Ernst Lindoorn and Jelle Westbroek for their help on the experiments setup. Thanks to all the friends who participated in the construction of the authentic emotion database. I would also like to Dr. Andre Deutz, Miss Riet Derogee and Miss Rachel Van Der Waal for helping me out with a problem on the scheduling of my oral presentation of master thesis. Finally a special word of thanks to Dr. Erwin Bakker for attending my graduation exam ceremony. 2 Neural Networks for Emotion Classification Table of Contents ABSTRACT ............................................................................................................................................5 1 INTRODUCTION ..........................................................................................................................5 2 BACKGROUND RESEARCH IN EMOTION RECOGNITION..............................................8 2.1 AN IDEAL SYSTEM FOR FACIAL EXPRESSION RECOGNITION: CURRENT PROBLEMS.....11 2.2 FACE DETECTION AND FEATURE EXTRACTION................................................................11 -
Truth Be Told
Truth Be Told . - washingtonpost.com Page 1 of 4 Hello vidal Change Preferences | Sign Out Print Edition | Subscribe NEWS POLITICS OPINIONS LOCAL SPORTS ARTS & LIVING CITY GUIDE JOBS SHO SEARCH: washingtonpost.com Web | Search Ar washingtonpost.com > Print Edition > Sunday Source » THIS STORY: READ + | Comments Truth Be Told . Sunday, November 25, 2007; Page N06 In the arena of lie detecting, it's important to remember that no emotion tells you its source. Think of Othello suffocating Desdemona because he interpreted her tears (over Cassio's death) as the reaction of an adulterer, not a friend. Othello made an assumption and killed his wife. Oops. The moral of the story: Just because someone exhibits the behavior of a liar does not make them one. "There isn't a silver bullet," psychologist Paul Ekman says. "That's because we don't have Pinocchio's nose. There is nothing in demeanor that is specific to a lie." Nevertheless, there are indicators that should prompt you to make further inquiries that may lead to discovering a lie. Below is a template for taking the first steps, gleaned from interviews with Ekman and several other experts, as well as their works. They are psychology professors Maureen O'Sullivan (University of San Francisco), (ISTOCKPHOTO) Robert Feldman (University of Massachusetts) and Bella DePaulo (University of California at Santa Barbara); TOOLBOX communication professor Mark Frank (University at Resize Text Save/Share + Buffalo); and body language trainer Janine Driver (a.k.a. Print This the Lyin' Tamer). E-mail This COMMENT How to Detect a Lie No comments have been posted about this http://www.washingtonpost.com/wp-dyn/content/article/2007/11/21/AR2007112102027... -
Ekman, Emotional Expression, and the Art of Empirical Epiphany
JOURNAL OF RESEARCH IN PERSONALITY Journal of Research in Personality 38 (2004) 37–44 www.elsevier.com/locate/jrp Ekman, emotional expression, and the art of empirical epiphany Dacher Keltner* Department of Psychology, University of California, Berkeley, 3319 Tolman, 94720 Berkeley, CA, USA Introduction In the mid and late 1960s, Paul Ekman offered a variety of bold assertions, some seemingly more radical today than others (Ekman, 1984, 1992, 1993). Emotions are expressed in a limited number of particular facial expressions. These expressions are universal and evolved. Facial expressions of emotion are remarkably brief, typically lasting 1 to 5 s. And germane to the interests of the present article, these brief facial expressions of emotion reveal a great deal about peopleÕs lives. In the present article I will present evidence that supports this last notion ad- vanced by Ekman, that brief expressions of emotion reveal important things about the individualÕs life course. To do so I first theorize about how individual differences in emotion shape the life context. With this reasoning as backdrop, I then review four kinds of evidence that indicate that facial expression is revealing of the life that the individual has led and is likely to continue leading. Individual differences in emotion and the shaping of the life context People, as a function of their personality or psychological disorder, create the sit- uations in which they act (e.g., Buss, 1987). Individuals selectively attend to certain features of complex situations, thus endowing contexts with idiosyncratic meaning. Individuals evoke responses in others, thus shaping the shared, social meaning of the situation. -
1 Automated Face Analysis for Affective Computing Jeffrey F. Cohn & Fernando De La Torre Abstract Facial Expression
Please do not quote. In press, Handbook of affective computing. New York, NY: Oxford Automated Face Analysis for Affective Computing Jeffrey F. Cohn & Fernando De la Torre Abstract Facial expression communicates emotion, intention, and physical state, and regulates interpersonal behavior. Automated Face Analysis (AFA) for detection, synthesis, and understanding of facial expression is a vital focus of basic research. While open research questions remain, the field has become sufficiently mature to support initial applications in a variety of areas. We review 1) human-observer based approaches to measurement that inform AFA; 2) advances in face detection and tracking, feature extraction, registration, and supervised learning; and 3) applications in action unit and intensity detection, physical pain, psychological distress and depression, detection of deception, interpersonal coordination, expression transfer, and other applications. We consider user-in-the-loop as well as fully automated systems and discuss open questions in basic and applied research. Keywords Automated Face Analysis and Synthesis, Facial Action Coding System (FACS), Continuous Measurement, Emotion, Nonverbal Communication, Synchrony 1. Introduction The face conveys information about a person’s age, sex, background, and identity, what they are feeling, or thinking (Darwin, 1872/1998; Ekman & Rosenberg, 2005). Facial expression regulates face-to-face interactions, indicates reciprocity and interpersonal attraction or repulsion, and enables inter-subjectivity between members of different cultures (Bråten, 2006; Fridlund, 1994; Tronick, 1989). Facial expression reveals comparative evolution, social and emotional development, neurological and psychiatric functioning, and personality processes (Burrows & Cohn, In press; Campos, Barrett, Lamb, Goldsmith, & Stenberg, 1983; Girard, Cohn, Mahoor, Mavadati, & Rosenwald, 2013; Schmidt & Cohn, 2001). Not surprisingly, the face has been of keen interest to behavioral scientists. -
E Motions in Process Barbara Rauch
e_motions in process Barbara Rauch Abstract This research project maps virtual emotions. Rauch uses 3D-surface capturing devices to scan facial expressions in (stuffed) animals and humans, which she then sculpts with the Phantom Arm/ SensAble FreeForm device in 3D virtual space. The results are rapidform printed objects and 3D animations of morphing faces and gestures. Building on her research into consciousness studies and emotions, she has developed a new artwork to reveal characteristic aspects of human emotions (i.e. laughing, crying, frowning, sneering, etc.), which utilises new technology, in particular digital scanning devices and special effects animation software. The proposal is to use a 3D high-resolution laser scanner to capture animal faces and, using the data of these faces, animate and then combine them with human emotional facial expressions. The morphing of the human and animal facial data are not merely layers of the different scans but by applying an algorithmic programme to the data, crucial landmarks in the animal face are merged in order to match with those of the human. The results are morphings of the physical characteristics of animals with the emotional characteristics of the human face in 3D. The focus of this interdisciplinary research project is a collaborative practice that brings together researchers from UCL in London and researchers at OCAD University’s data and information visualization lab. Rauch uses Darwin’s metatheory of the continuity of species and other theories on evolution and internal physiology (Ekman et al) in order to re-examine previous and new theories with the use of new technologies, including the SensAble FreeForm Device, which, as an interface, allows for haptic feedback from digital data.Keywords: interdisciplinary research, 3D-surface capturing, animated facial expressions, evolution of emotions and feelings, technologically transformed realities. -
Facial Expression and Emotion Paul Ekman
1992 Award Addresses Facial Expression and Emotion Paul Ekman Cross-cultural research on facial expression and the de- We found evidence of universality in spontaneous velopments of methods to measure facial expression are expressions and in expressions that were deliberately briefly summarized. What has been learned about emo- posed. We postulated display rules—culture-specific pre- tion from this work on the face is then elucidated. Four scriptions about who can show which emotions, to whom, questions about facial expression and emotion are dis- and when—to explain how cultural differences may con- cussed: What information does an expression typically ceal universal in expression, and in an experiment we convey? Can there be emotion without facial expression? showed how that could occur. Can there be a facial expression of emotion without emo- In the last five years, there have been a few challenges tion? How do individuals differ in their facial expressions to the evidence of universals, particularly from anthro- of emotion? pologists (see review by Lutz & White, 1986). There is. however, no quantitative data to support the claim that expressions are culture specific. The accounts are more In 1965 when I began to study facial expression,1 few- anecdotal, without control for the possibility of observer thought there was much to be learned. Goldstein {1981} bias and without evidence of interobserver reliability. pointed out that a number of famous psychologists—F. There have been recent challenges also from psychologists and G. Allport, Brunswik, Hull, Lindzey, Maslow, Os- (J. A. Russell, personal communication, June 1992) who good, Titchner—did only one facial study, which was not study how words are used to judge photographs of facial what earned them their reputations. -
Emotion Classification Based on Biophysical Signals and Machine Learning Techniques
S S symmetry Article Emotion Classification Based on Biophysical Signals and Machine Learning Techniques Oana Bălan 1,* , Gabriela Moise 2 , Livia Petrescu 3 , Alin Moldoveanu 1 , Marius Leordeanu 1 and Florica Moldoveanu 1 1 Faculty of Automatic Control and Computers, University POLITEHNICA of Bucharest, Bucharest 060042, Romania; [email protected] (A.M.); [email protected] (M.L.); fl[email protected] (F.M.) 2 Department of Computer Science, Information Technology, Mathematics and Physics (ITIMF), Petroleum-Gas University of Ploiesti, Ploiesti 100680, Romania; [email protected] 3 Faculty of Biology, University of Bucharest, Bucharest 030014, Romania; [email protected] * Correspondence: [email protected]; Tel.: +40722276571 Received: 12 November 2019; Accepted: 18 December 2019; Published: 20 December 2019 Abstract: Emotions constitute an indispensable component of our everyday life. They consist of conscious mental reactions towards objects or situations and are associated with various physiological, behavioral, and cognitive changes. In this paper, we propose a comparative analysis between different machine learning and deep learning techniques, with and without feature selection, for binarily classifying the six basic emotions, namely anger, disgust, fear, joy, sadness, and surprise, into two symmetrical categorical classes (emotion and no emotion), using the physiological recordings and subjective ratings of valence, arousal, and dominance from the DEAP (Dataset for Emotion Analysis using EEG, Physiological and Video Signals) database. The results showed that the maximum classification accuracies for each emotion were: anger: 98.02%, joy:100%, surprise: 96%, disgust: 95%, fear: 90.75%, and sadness: 90.08%. In the case of four emotions (anger, disgust, fear, and sadness), the classification accuracies were higher without feature selection. -
Conceptual Framework for Quantum Affective Computing and Its Use in Fusion of Multi-Robot Emotions
electronics Article Conceptual Framework for Quantum Affective Computing and Its Use in Fusion of Multi-Robot Emotions Fei Yan 1 , Abdullah M. Iliyasu 2,3,∗ and Kaoru Hirota 3,4 1 School of Computer Science and Technology, Changchun University of Science and Technology, Changchun 130022, China; [email protected] 2 College of Engineering, Prince Sattam Bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia 3 School of Computing, Tokyo Institute of Technology, Yokohama 226-8502, Japan; [email protected] 4 School of Automation, Beijing Institute of Technology, Beijing 100081, China * Correspondence: [email protected] Abstract: This study presents a modest attempt to interpret, formulate, and manipulate the emotion of robots within the precepts of quantum mechanics. Our proposed framework encodes emotion information as a superposition state, whilst unitary operators are used to manipulate the transition of emotion states which are subsequently recovered via appropriate quantum measurement operations. The framework described provides essential steps towards exploiting the potency of quantum mechanics in a quantum affective computing paradigm. Further, the emotions of multi-robots in a specified communication scenario are fused using quantum entanglement, thereby reducing the number of qubits required to capture the emotion states of all the robots in the environment, and therefore fewer quantum gates are needed to transform the emotion of all or part of the robots from one state to another. In addition to the mathematical rigours expected of the proposed framework, we present a few simulation-based demonstrations to illustrate its feasibility and effectiveness. This exposition is an important step in the transition of formulations of emotional intelligence to the quantum era. -
Optimal Arousal Identification and Classification for Affective Computing Using Physiological Signals: Virtual Reality Stroop Task
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, VOL. 1, NO. 2, JULY-DECEMBER 2010 109 Optimal Arousal Identification and Classification for Affective Computing Using Physiological Signals: Virtual Reality Stroop Task Dongrui Wu, Member, IEEE, Christopher G. Courtney, Brent J. Lance, Member, IEEE, Shrikanth S. Narayanan, Fellow, IEEE, Michael E. Dawson, Kelvin S. Oie, and Thomas D. Parsons Abstract—A closed-loop system that offers real-time assessment and manipulation of a user’s affective and cognitive states is very useful in developing adaptive environments which respond in a rational and strategic fashion to real-time changes in user affect, cognition, and motivation. The goal is to progress the user from suboptimal cognitive and affective states toward an optimal state that enhances user performance. In order to achieve this, there is need for assessment of both 1) the optimal affective/cognitive state and 2) the observed user state. This paper presents approaches for assessing these two states. Arousal, an important dimension of affect, is focused upon because of its close relation to a user’s cognitive performance, as indicated by the Yerkes-Dodson Law. Herein, we make use of a Virtual Reality Stroop Task (VRST) from the Virtual Reality Cognitive Performance Assessment Test (VRCPAT) to identify the optimal arousal level that can serve as the affective/cognitive state goal. Three stimuli presentations (with distinct arousal levels) in the VRST are selected. We demonstrate that when reaction time is used as the performance measure, one of the three stimuli presentations can elicit the optimal level of arousal for most subjects. Further, results suggest that high classification rates can be achieved when a support vector machine is used to classify the psychophysiological responses (skin conductance level, respiration, ECG, and EEG) in these three stimuli presentations into three arousal levels. -
A Fuzzy Modelling Approach of Emotion for Affective Computing Systems
A Fuzzy Modelling Approach of Emotion for Affective Computing Systems Charalampos Karyotis1, Faiyaz Doctor1, Rahat Iqbal1, Anne James1 and Victor Chang2 1Faculty of Engineering, Environment and Computing, Coventry University, Priory Street, CV1 5FB, Coventry, U.K. 2School of Computing, Creative Technologies & Engineering, Leeds Beckett University, City Campus, Leeds, LS1 3HE, U.K. Keywords: Adaptive Fuzzy Systems, Emotion Modelling, Affective Trajectories, Arousal Valence, Affective Computing, Personalised Learning. Abstract: In this paper we present a novel affective modelling approach to be utilised by Affective Computing systems. This approach is a combination of the well known Arousal Valence model of emotion and the newly introduced Affective Trajectories Hypothesis. An adaptive data driven fuzzy method is proposed in order to extract personalized emotion models, and successfully visualise the associations of these models’ basic elements, to different emotional labels, using easily interpretable fuzzy rules. Namely we explore how the combinations of arousal, valence, prediction of the future, and the experienced outcome after this prediction, enable us to differentiate between different emotional labels. We use the results obtained from a user study consisting of an online survey, to demonstrate the potential applicability of this affective modelling approach, and test the effectiveness and stability of its adaptive element, which accounts for individual differences between the users. We also propose a basic architecture in order for this approach to be used effectively by AC systems, and finally we present an implementation of a personalised learning system which utilises the suggested framework. This implementation is tested through a pilot experimental session consisting of a tutorial on fuzzy logic which was conducted under an activity-led and problem based learning context. -
Facial Expressions of Emotion Influence Interpersonal Trait Inferences
FACIAL EXPRESSIONS OF EMOTION INFLUENCE INTERPERSONAL TRAIT INFERENCES Brian Knutson ABSTRACT.. Theorists have argued that facial expressions of emotion serve the inter- personal function of allowing one animal to predict another's behavior. Humans may extend these predictions into the indefinite future, as in the case of trait infer- ence. The hypothesis that facial expressions of emotion (e.g., anger, disgust, fear, happiness, and sadness) affect subjects' interpersonal trait inferences (e.g., domi- nance and affiliation) was tested in two experiments. Subjects rated the disposi- tional affiliation and dominance of target faces with either static or apparently mov- ing expressions. They inferred high dominance and affiliation from happy expressions, high dominance and low affiliation from angry and disgusted expres- sions, and low dominance from fearful and sad expressions. The findings suggest that facial expressions of emotion convey not only a targets internal state, but also differentially convey interpersonal information, which could potentially seed trait inference. What do people infer from facial expressions of emotion? Darwin (1872/1962) suggested that facial muscle movements which originally sub- served individual survival problems (e.g., spitting out noxious food, shield- ing the eyes) eventually allowed animals to predict the behavior of their conspecifics. Current emotion theorists agree that emotional facial expres- sions can serve social predictive functions (e.g., Ekman, 1982, Izard, 1972; Plutchik, 1980). For instance, Frank (1988) hypothesizes that a person might signal a desire to cooperate by smiling at another. Although a viewer may predict a target's immediate behavior on the basis of his or her facial expressions, the viewer may also extrapolate to the more distant future, as in the case of inferring personality traits. -
An Affective Computing Approach
Assessing Emotions by Cursor Motions: An Affective Computing Approach Takashi Yamauchi 1, Hwaryong Seo 2, Yoonsuck Choe 3, Casady Bowman 1, and Kunchen Xiao 1 1Department of Psychology, 2Department of Visualization, 3Department of Computer Science and Engineering Texas A&M University, TX 77843 ([email protected]) Abstract technologies: (1) many of the visual- and audio-based methods (e.g., detecting emotions by facial expressions and Choice reaching, e.g., reaching a targeted object by hand, involves a dynamic online integration of perception, action and speech) do not fare well in a natural setting; (2) assessment cognition, where neural activities of prefrontal cortical regions methods based on physiological signals (e.g., EEG) are still are concurrently coordinated with sensori-motor subsystems. impractical for everyday application. Our independent On the basis of this theoretical development, the authors review of the studies published in major Human Computer investigate the extent to which cursor movements in a simple Interaction (HCI) conferences and journals show significant choice-reaching task reveal people’s emotions, such as anxiety. improvements in AC technologies in the last several years. The results show that there is a strong correlation between cursor trajectory patterns and self-reported anxiety in male Techniques developed in “wearable computers” made great participants. Because computer cursors are ubiquitous, our progress in assessing people’s physiological states in trajectory analysis can be augmented to