Formulating Emotion Perception As a Probabilistic Model with Application to Categorical Emotion Classification

Total Page:16

File Type:pdf, Size:1020Kb

Formulating Emotion Perception As a Probabilistic Model with Application to Categorical Emotion Classification 2017 Seventh International Conference on Affective Computing and Intelligent Interaction (ACII) Formulating Emotion Perception as a Probabilistic Model with Application to Categorical Emotion Classification Reza Lotfian Carlos Busso The University of Texas at Dallas The University of Texas at Dallas Email: [email protected] Email: [email protected] Abstract—Automatic recognition of emotions is an important sample. Then, a consensus labels, such as majority vote, part of affect-sensitive human-computer interaction (HCI). Ex- is calculated and assigned as ground truth. This approach pressive behaviors tend to be ambiguous with blended emotions aims to identify popular trends where outlier assessments during natural spontaneous conversations. Therefore, evalu- are ignored. While annotators may report conflicting cat- ators disagree on the perceived emotion, assigning multiple egories, all of them might be right, given the differences emotional classes to the same stimuli (e.g., sadness, anger, sur- in perception and ambiguity in the expressed emotions. prise). These observations have clear implications on emotion Emotions are driven by appraisal of a given situation [5]. classification, where assigning a single descriptor per stimuli Depending on the evaluator’s perspective, multiple answers oversimplifies the intrinsic subjectivity in emotion perception. may be appropriate, especially if many related emotional This study proposes a new formulation, where the emotional categories are available (e.g., surprise, anger, fear, disgust). perception of a stimuli is a multidimensional Gaussian random We hypothesize that individual evaluations provide richer variable with an unobserved distribution. Each dimension emotional descriptions than simple consensus labels, which corresponds to an emotion characterized by a numerical scale. should be effectively leveraged while training classifiers [6], The covariance matrix of this distribution captures the intrin- [7], [8]. The formulation of the classifier should understand sic dependencies between different emotional categories. The the relationship between emotional classes, prioritizing the process where an evaluator judges the stimuli is equivalent to separation of unrelated categories (e.g., anger versus happi- ness) over related emotions (anger versus disgust) [9]. sampling a point from this distribution, reporting the class with the highest value. The proposed approach recursively estimates This study proposes an innovative formulation to lever- this multimodal distribution using numerical methods. The age individual evaluations for emotion recognition, as op- mean of the Gaussian distribution is used as a soft label to posed to consensus labels, considering the intrinsic relation- train a deep neural network (DNN). Our experimental results ship between emotional categories. We formulate emotion show that the proposed training method leads to improvements perception as a probabilistic model, where each speech in F-score over training with (1) hard-labels based on majority segment has a non-observable multivariate Gaussian dis- vote, and (2) soft-label framework proposed by other studies. tribution. Each dimension is associated with the intensity measure of one emotion category (i.e., dimension equals the 1. Introduction number of emotions). Emotional perception is equivalent to Recognizing categorical emotions such as happiness, sampling a point according to this multivariate distribution. sadness or anger from speech can have many practical After listening to the stimuli, the evaluator draws a point applications [1], [2], [3]. Previous studies on categorical in the distribution, reporting the emotional category that emotion recognition rely on the assumption that (1) each has the largest intensity measure. Each individual evaluation speech segment is often assigned to one emotional class, corresponds to a realization drawn from this distribution, and (2) samples belonging to the same class share similar and the task is to approximate this multivariate Gaussian acoustic features. However, the boundaries between emotion distribution using numerical methods. We demonstrate that classes during human interaction are ambiguous, where peo- this formulation can be used in emotion recognition by (1) ple rarely express extreme emotions. Samples labeled with training the classifiers using the expected value of each emo- a given emotional class may have different acoustic char- tional dimension, derived from random Gaussian distribution acteristics (e.g., shades of happiness). These observations of individual samples (e.g., soft labels), and (2) modifying have direct implication in machine learning frameworks for the loss function to weigh errors between emotional class emotion recognition, where conventional approaches do not according to their relations. We refer to this method as soft generalize when evaluated in realistic scenarios [4]. label from the expected intensity of emotion (SL-EIE). The The emotional annotation process of spontaneous proposed method achieves better performance than classi- databases often include perceptual evaluations, where sev- fiers trained with hard labels derived from majority vote eral evaluators are asked to assign an emotional class to each and soft-label proposed by Fayek et al. [7]. 978-1-5386-0563-9/17/$31.00 c 2017 IEEE 415 2. Related work Unlike other speech processing problems such as speaker identification or automatic speech recognition (ASR), defining ground-truth labels for speech emotion recognition is not straightforward. Emotions observed dur- ing human interactions are ambiguous, which are not nec- essarily captured by the target emotional descriptions or the annotation protocol (e.g., forcing an evaluator to choose a class from a predefined list [10]). The standard approach is to collect multiple annotations from experts or na¨ıve eval- uators, aiming to reach a consensus label. Given subjective nature of perceptual evaluations, the resulting ground-truth labels are noisy [9], [11]. Figure 1. Illustration of approach to formulate emotion perception as a When the labels are inherently noisy, studies have pro- probabilistic process. Unobservable distributions of the intensity of happi- ness and neutral for two sentences, evaluated by four evaluators. posed using ensembles [12] and soft-labels [13], describ- ing the probability or intensity associated with each class. sample. The evaluator selects the class with higher intensity In speech emotion recognition, studies have proposed dif- according to his/her judgement. Cases when two annotators ferent variants of this approach, leveraging the individual provide conflicting emotions for a speech segment may evaluations provided to each speech sample (i.e., not just indicate that they perceive emotions with different intensity the consensus label). Mower et al. [14], [15] suggested level. Therefore, the selected class does not imply complete emotional profiles to describe the confidence that an emo- disagreement. We propose to estimate soft label from the tional label is assigned to an utterance. Audhkhasi and expected intensity of emotion (SL-EIE), which we estimate Narayanan [8] proposed fuzzy logic to deal with ambiguous from the individual evaluations. emotional utterances by generalizing class sets with partial membership, rather than single class membership. The study We formulate the problem by modeling perceptual eval- suggested that more information can be extracted from uation as a random vector, where each dimension corre- individual annotations assigned to a given speech sample sponds to one emotional class. The multivariate distribution than a consensus label. Other studies have also explored of this process is hidden and the task is to estimate it the use of individual annotations to create relative labels from the individual evaluations. Figure 1 illustrates the happiness of categorical emotions for preference learning [6]. The key ideas when we consider only two emotions ( neutral study derived a probabilistic relevant score from individual and ) for two speech segments (red and blue). The x neutral annotations, indicating the intensity of a given emotional -axis corresponds to the perceived intensity for , y class (e.g., intensity of happiness in a speech segment). The and the -axis corresponds to the perceived intensity for happiness approach was successfully used to train emotional rankers. Each segment is evaluated by four annotators. Fayek et al. [7] used individual evaluations to create This process is formulated as drawing four points from ensembles with deep neural network (DNN). Each DNN each of these distributions (red points for sentence 1, blue is trained with the annotations from one annotator, pro- squares for sentence 2). Based on the values for x and viding the final score after fusing the ensembles. Alter- y, the evaluator chooses neutral if x > y or happiness if natively, the study trained a single DNN with soft labels y > x. The resulting labels for these eight evaluations are reflecting the proportion of evaluators selecting each class. denoted with H for happiness and N for neutral. Notice Their experimental results show the benefit of considering that the line y = x defines the boundaries between the individual evaluations in building emotion classifiers. This emotions. Majority vote will suggest that one sample is study is related to our paper, so we use this approach happy and the other is neutral. Instead, we aim to estimate as one of the baselines. Our proposed method
Recommended publications
  • Improving Emotion Perception and Emotion Regulation Through a Web-Based Emotional Intelligence Training (WEIT) Program for Future Leaders
    Volume 11, Number 2, November 2019 pp 17 - 32 www.um.edu.mt/ijee Improving Emotion Perception and Emotion Regulation Through a Web-Based Emotional Intelligence Training (WEIT) Program for Future Leaders 1 Christina Köppe, Marco Jürgen Held and Astrid Schütz University of Bamberg, Bamberg, Germany We evaluated a Web-Based Emotional Intelligence Training (WEIT) program that was based on the four-branch model of emotional intelligence (EI) and which aimed at improving emotion perception (EP) and emotion regulation (ER) in future leaders. Using a controlled experimental design, we evaluated the short-term (directly after the WEIT program) and long-term (6 weeks later) effects in a sample of 134 (59 training group [TG], 75 wait list control group [CG]) business students, and additionally tested whether WEIT helped to reduce perceived stress. For EP, WEIT led to a significant increase in the TG directly after training (whereas the wait list CG showed no change). Changes remained stable after 6 weeks in the TG, but there were no significant differences between the TG and CG at follow-up. By contrast, ER did not show an increase directly after WEIT, but 6 weeks later, the TG had larger improvements than the CG. The results mostly confirmed that emotional abilities can be increased through web-based training. Participants’ perceived stress did not decrease after the training program. Further refinement and validation of WEIT is needed. Keywords: emotional intelligence, web-based training, stress, future leaders First submission 15th May 2019; Accepted for publication 4th September 2019. Introduction Emotional intelligence (EI) has attracted considerable attention in recent years (see Côté, 2014).
    [Show full text]
  • Emotion Detection in Twitter Posts: a Rule-Based Algorithm for Annotated Data Acquisition
    2020 International Conference on Computational Science and Computational Intelligence (CSCI) Emotion detection in Twitter posts: a rule-based algorithm for annotated data acquisition Maria Krommyda Anastatios Rigos Institute of Communication and Computer Systems Institute of Communication and Computer Systems Athens, Greece Athens, Greece [email protected] [email protected] Kostas Bouklas Angelos Amditis Institute of Communication and Computer Systems Institute of Communication and Computer Systems Athens, Greece Athens, Greece [email protected] [email protected] Abstract—Social media analysis plays a key role to the person that provided the text and it is categorized as positive, understanding of the public’s opinion regarding recent events negative, or neutral. Such analysis is mainly used for movies, and decisions, to the design and management of advertising products and persons, when measuring their appeal to the campaigns as well as to the planning of next steps and mitigation actions for public relationship initiatives. Significant effort has public is of interest [4] and require extended quantities of text been dedicated recently to the development of data analysis collected over a long period of time from different sources to algorithms that will perform, in an automated way, sentiment ensure an unbiased and objective output. analysis over publicly available text. Most of the available While this technique is very popular and well established, research work, focuses on binary categorizing text as positive or with many important use cases and applications, there are negative without further investigating the emotions leading to that categorization. The current needs, however, for in-depth analysis other equally important use cases where such analysis fail to of the available content combined with the complexity and multi- provided the needed information.
    [Show full text]
  • Emotion Perception in Habitual Players of Action Video Games Swann Pichon, Benoit Bediou, Lia Antico, Rachael Jack, Oliver Garrod, Chris Sims, C
    Emotion Emotion Perception in Habitual Players of Action Video Games Swann Pichon, Benoit Bediou, Lia Antico, Rachael Jack, Oliver Garrod, Chris Sims, C. Shawn Green, Philippe Schyns, and Daphne Bavelier Online First Publication, July 6, 2020. http://dx.doi.org/10.1037/emo0000740 CITATION Pichon, S., Bediou, B., Antico, L., Jack, R., Garrod, O., Sims, C., Green, C. S., Schyns, P., & Bavelier, D. (2020, July 6). Emotion Perception in Habitual Players of Action Video Games. Emotion. Advance online publication. http://dx.doi.org/10.1037/emo0000740 Emotion © 2020 American Psychological Association 2020, Vol. 2, No. 999, 000 ISSN: 1528-3542 http://dx.doi.org/10.1037/emo0000740 Emotion Perception in Habitual Players of Action Video Games Swann Pichon, Benoit Bediou, and Lia Antico Rachael Jack and Oliver Garrod University of Geneva, Campus Biotech Glasgow University Chris Sims C. Shawn Green Rensselaer Polytechnic Institute University of Wisconsin–Madison Philippe Schyns Daphne Bavelier Glasgow University University of Geneva, Campus Biotech Action video game players (AVGPs) display superior performance in various aspects of cognition, especially in perception and top-down attention. The existing literature has examined these performance almost exclusively with stimuli and tasks devoid of any emotional content. Thus, whether the superior performance documented in the cognitive domain extend to the emotional domain remains unknown. We present 2 cross-sectional studies contrasting AVGPs and nonvideo game players (NVGPs) in their ability to perceive facial emotions. Under an enhanced perception account, AVGPs should outperform NVGPs when processing facial emotion. Yet, alternative accounts exist. For instance, under some social accounts, exposure to action video games, which often contain violence, may lower sensitivity for empathy-related expressions such as sadness, happiness, and pain while increasing sensitivity to aggression signals.
    [Show full text]
  • 1 Automated Face Analysis for Affective Computing Jeffrey F. Cohn & Fernando De La Torre Abstract Facial Expression
    Please do not quote. In press, Handbook of affective computing. New York, NY: Oxford Automated Face Analysis for Affective Computing Jeffrey F. Cohn & Fernando De la Torre Abstract Facial expression communicates emotion, intention, and physical state, and regulates interpersonal behavior. Automated Face Analysis (AFA) for detection, synthesis, and understanding of facial expression is a vital focus of basic research. While open research questions remain, the field has become sufficiently mature to support initial applications in a variety of areas. We review 1) human-observer based approaches to measurement that inform AFA; 2) advances in face detection and tracking, feature extraction, registration, and supervised learning; and 3) applications in action unit and intensity detection, physical pain, psychological distress and depression, detection of deception, interpersonal coordination, expression transfer, and other applications. We consider user-in-the-loop as well as fully automated systems and discuss open questions in basic and applied research. Keywords Automated Face Analysis and Synthesis, Facial Action Coding System (FACS), Continuous Measurement, Emotion, Nonverbal Communication, Synchrony 1. Introduction The face conveys information about a person’s age, sex, background, and identity, what they are feeling, or thinking (Darwin, 1872/1998; Ekman & Rosenberg, 2005). Facial expression regulates face-to-face interactions, indicates reciprocity and interpersonal attraction or repulsion, and enables inter-subjectivity between members of different cultures (Bråten, 2006; Fridlund, 1994; Tronick, 1989). Facial expression reveals comparative evolution, social and emotional development, neurological and psychiatric functioning, and personality processes (Burrows & Cohn, In press; Campos, Barrett, Lamb, Goldsmith, & Stenberg, 1983; Girard, Cohn, Mahoor, Mavadati, & Rosenwald, 2013; Schmidt & Cohn, 2001). Not surprisingly, the face has been of keen interest to behavioral scientists.
    [Show full text]
  • A Review of Alexithymia and Emotion Perception in Music, Odor, Taste, and Touch
    MINI REVIEW published: 30 July 2021 doi: 10.3389/fpsyg.2021.707599 Beyond Face and Voice: A Review of Alexithymia and Emotion Perception in Music, Odor, Taste, and Touch Thomas Suslow* and Anette Kersting Department of Psychosomatic Medicine and Psychotherapy, University of Leipzig Medical Center, Leipzig, Germany Alexithymia is a clinically relevant personality trait characterized by deficits in recognizing and verbalizing one’s emotions. It has been shown that alexithymia is related to an impaired perception of external emotional stimuli, but previous research focused on emotion perception from faces and voices. Since sensory modalities represent rather distinct input channels it is important to know whether alexithymia also affects emotion perception in other modalities and expressive domains. The objective of our review was to summarize and systematically assess the literature on the impact of alexithymia on the perception of emotional (or hedonic) stimuli in music, odor, taste, and touch. Eleven relevant studies were identified. On the basis of the reviewed research, it can be preliminary concluded that alexithymia might be associated with deficits Edited by: in the perception of primarily negative but also positive emotions in music and a Mathias Weymar, University of Potsdam, Germany reduced perception of aversive taste. The data available on olfaction and touch are Reviewed by: inconsistent or ambiguous and do not allow to draw conclusions. Future investigations Khatereh Borhani, would benefit from a multimethod assessment of alexithymia and control of negative Shahid Beheshti University, Iran Kristen Paula Morie, affect. Multimodal research seems necessary to advance our understanding of emotion Yale University, United States perception deficits in alexithymia and clarify the contribution of modality-specific and Jan Terock, supramodal processing impairments.
    [Show full text]
  • Machine Learning Based Emotion Classification Using the Covid-19 Real World Worry Dataset
    Anatolian Journal of Computer Sciences Volume:6 No:1 2021 © Anatolian Science pp:24-31 ISSN:2548-1304 Machine Learning Based Emotion Classification Using The Covid-19 Real World Worry Dataset Hakan ÇAKAR1*, Abdulkadir ŞENGÜR2 *1Fırat Üniversitesi/Teknoloji Fak., Elektrik-Elektronik Müh. Bölümü, Elazığ, Türkiye ([email protected]) 2Fırat Üniversitesi/Teknoloji Fak., Elektrik-Elektronik Müh. Bölümü, Elazığ, Türkiye ([email protected]) Received Date : Sep. 27, 2020 Acceptance Date : Jan. 1, 2021 Published Date : Mar. 1, 2021 Özetçe— COVID-19 pandemic has a dramatic impact on economies and communities all around the world. With social distancing in place and various measures of lockdowns, it becomes significant to understand emotional responses on a great scale. In this paper, a study is presented that determines human emotions during COVID-19 using various machine learning (ML) approaches. To this end, various techniques such as Decision Trees (DT), Support Vector Machines (SVM), k-nearest neighbor (k-NN), Neural Networks (NN) and Naïve Bayes (NB) methods are used in determination of the human emotions. The mentioned techniques are used on a dataset namely Real World Worry dataset (RWWD) that was collected during COVID-19. The dataset, which covers eight emotions on a 9-point scale, grading their anxiety levels about the COVID-19 situation, was collected by using 2500 participants. The performance evaluation of the ML techniques on emotion prediction is carried out by using the accuracy score. Five-fold cross validation technique is also adopted in experiments. The experiment works show that the ML approaches are promising in determining the emotion in COVID- 19 RWWD.
    [Show full text]
  • Emotion Classification Based on Biophysical Signals and Machine Learning Techniques
    S S symmetry Article Emotion Classification Based on Biophysical Signals and Machine Learning Techniques Oana Bălan 1,* , Gabriela Moise 2 , Livia Petrescu 3 , Alin Moldoveanu 1 , Marius Leordeanu 1 and Florica Moldoveanu 1 1 Faculty of Automatic Control and Computers, University POLITEHNICA of Bucharest, Bucharest 060042, Romania; [email protected] (A.M.); [email protected] (M.L.); fl[email protected] (F.M.) 2 Department of Computer Science, Information Technology, Mathematics and Physics (ITIMF), Petroleum-Gas University of Ploiesti, Ploiesti 100680, Romania; [email protected] 3 Faculty of Biology, University of Bucharest, Bucharest 030014, Romania; [email protected] * Correspondence: [email protected]; Tel.: +40722276571 Received: 12 November 2019; Accepted: 18 December 2019; Published: 20 December 2019 Abstract: Emotions constitute an indispensable component of our everyday life. They consist of conscious mental reactions towards objects or situations and are associated with various physiological, behavioral, and cognitive changes. In this paper, we propose a comparative analysis between different machine learning and deep learning techniques, with and without feature selection, for binarily classifying the six basic emotions, namely anger, disgust, fear, joy, sadness, and surprise, into two symmetrical categorical classes (emotion and no emotion), using the physiological recordings and subjective ratings of valence, arousal, and dominance from the DEAP (Dataset for Emotion Analysis using EEG, Physiological and Video Signals) database. The results showed that the maximum classification accuracies for each emotion were: anger: 98.02%, joy:100%, surprise: 96%, disgust: 95%, fear: 90.75%, and sadness: 90.08%. In the case of four emotions (anger, disgust, fear, and sadness), the classification accuracies were higher without feature selection.
    [Show full text]
  • DEAP: a Database for Emotion Analysis Using Physiological Signals
    IEEE TRANS. AFFECTIVE COMPUTING 1 DEAP: A Database for Emotion Analysis using Physiological Signals Sander Koelstra, Student Member, IEEE, Christian M¨uhl, Mohammad Soleymani, Student Member, IEEE, Jong-Seok Lee, Member, IEEE, Ashkan Yazdani, Touradj Ebrahimi, Member, IEEE, Thierry Pun, Member, IEEE, Anton Nijholt, Member, IEEE, Ioannis Patras, Member, IEEE Abstract—We present a multimodal dataset for the analysis of human affective states. The electroencephalogram (EEG) and peripheral physiological signals of 32 participants were recorded as each watched 40 one-minute long excerpts of music videos. Participants rated each video in terms of the levels of arousal, valence, like/dislike, dominance and familiarity. For 22 of the 32 participants, frontal face video was also recorded. A novel method for stimuli selection is proposed using retrieval by affective tags from the last.fm website, video highlight detection and an online assessment tool. An extensive analysis of the participants’ ratings during the experiment is presented. Correlates between the EEG signal frequencies and the participants’ ratings are investigated. Methods and results are presented for single-trial classification of arousal, valence and like/dislike ratings using the modalities of EEG, peripheral physiological signals and multimedia content analysis. Finally, decision fusion of the classification results from the different modalities is performed. The dataset is made publicly available and we encourage other researchers to use it for testing their own affective state estimation methods. Index Terms—Emotion classification, EEG, Physiological signals, Signal processing, Pattern classification, Affective computing. ✦ 1 INTRODUCTION information retrieval. Affective characteristics of multi- media are important features for describing multime- MOTION is a psycho-physiological process triggered dia content and can be presented by such emotional by conscious and/or unconscious perception of an E tags.
    [Show full text]
  • Emotion Classification Based on Brain Wave: a Survey
    Li et al. Hum. Cent. Comput. Inf. Sci. (2019) 9:42 https://doi.org/10.1186/s13673-019-0201-x REVIEW Open Access Emotion classifcation based on brain wave: a survey Ting‑Mei Li1, Han‑Chieh Chao1,2* and Jianming Zhang3 *Correspondence: [email protected] Abstract 1 Department of Electrical Brain wave emotion analysis is the most novel method of emotion analysis at present. Engineering, National Dong Hwa University, Hualien, With the progress of brain science, it is found that human emotions are produced by Taiwan the brain. As a result, many brain‑wave emotion related applications appear. However, Full list of author information the analysis of brain wave emotion improves the difculty of analysis because of the is available at the end of the article complexity of human emotion. Many researchers used diferent classifcation methods and proposed methods for the classifcation of brain wave emotions. In this paper, we investigate the existing methods of brain wave emotion classifcation and describe various classifcation methods. Keywords: EEG, Emotion, Classifcation Introduction Brain science has shown that human emotions are controlled by the brain [1]. Te brain produces brain waves as it transmits messages. Brain wave data is one of the biological messages, and biological messages usually have emotion features. Te feature of emo- tion can be extracted through the analysis of brain wave messages. But because of the environmental background and cultural diferences of human growth, the complexity of human emotion is caused. Terefore, in the analysis of brain wave emotion, classifcation algorithm is very important. In this article, we will focus on the classifcation of brain wave emotions.
    [Show full text]
  • EMOTION CLASSIFICATION of COVERED FACES 1 WHO CAN READ YOUR FACIAL EXPRESSION? a Comparison of Humans and Machine Learning Class
    journal name manuscript No. (will be inserted by the editor) 1 A Comparison of Humans and Machine Learning 2 Classifiers Detecting Emotion from Faces of People 3 with Different Coverings 4 Harisu Abdullahi Shehu · Will N. 5 Browne · Hedwig Eisenbarth 6 7 Received: 9th August 2021 / Accepted: date 8 Abstract Partial face coverings such as sunglasses and face masks uninten- 9 tionally obscure facial expressions, causing a loss of accuracy when humans 10 and computer systems attempt to categorise emotion. Experimental Psychol- 11 ogy would benefit from understanding how computational systems differ from 12 humans when categorising emotions as automated recognition systems become 13 available. We investigated the impact of sunglasses and different face masks on 14 the ability to categorize emotional facial expressions in humans and computer 15 systems to directly compare accuracy in a naturalistic context. Computer sys- 16 tems, represented by the VGG19 deep learning algorithm, and humans assessed 17 images of people with varying emotional facial expressions and with four dif- 18 ferent types of coverings, i.e. unmasked, with a mask covering lower-face, a 19 partial mask with transparent mouth window, and with sunglasses. Further- Harisu Abdullahi Shehu (Corresponding author) School of Engineering and Computer Science, Victoria University of Wellington, 6012 Wellington E-mail: [email protected] ORCID: 0000-0002-9689-3290 Will N. Browne School of Electrical Engineering and Robotics, Queensland University of Technology, Bris- bane QLD 4000, Australia Tel: +61 45 2468 148 E-mail: [email protected] ORCID: 0000-0001-8979-2224 Hedwig Eisenbarth School of Psychology, Victoria University of Wellington, 6012 Wellington Tel: +64 4 463 9541 E-mail: [email protected] ORCID: 0000-0002-0521-2630 2 Shehu H.
    [Show full text]
  • Hybrid Approach for Emotion Classification of Audio Conversation Based on Text and Speech Mining
    Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 46 ( 2015 ) 635 – 643 International Conference on Information and Communication Technologies (ICICT 2014) Hybrid Approach for Emotion Classification of Audio Conversation Based on Text and Speech Mining a, a b Jasmine Bhaskar * ,Sruthi K , Prema Nedungadi aDepartment of Compute science , Amrita University, Amrita School of Engineering, Amritapuri, 690525, India bAmrita Create, Amrita University, Amrita School of Engineering, Amritapuri, 690525, India Abstract One of the greatest challenges in speech technology is estimating the speaker’s emotion. Most of the existing approaches concentrate either on audio or text features. In this work, we propose a novel approach for emotion classification of audio conversation based on both speech and text. The novelty in this approach is in the choice of features and the generation of a single feature vector for classification. Our main intention is to increase the accuracy of emotion classification of speech by considering both audio and text features. In this work we use standard methods such as Natural Language Processing, Support Vector Machines, WordNet Affect and SentiWordNet. The dataset for this work have been taken from Semval -2007 and eNTERFACE'05 EMOTION Database. © 20152014 PublishedThe Authors. by ElsevierPublished B.V. by ElsevierThis is an B.V. open access article under the CC BY-NC-ND license (Peer-reviewhttp://creativecommons.org/licenses/by-nc-nd/4.0/ under responsibility of organizing committee). of the International Conference on Information and Communication PeerTechnologies-review under (ICICT responsibility 2014). of organizing committee of the International Conference on Information and Communication Technologies (ICICT 2014) Keywords: Emotion classification;text mining;speech mining;hybrid approach 1.
    [Show full text]
  • Postural Communication of Emotion: Perception of Distinct Poses of Five Discrete Emotions
    fpsyg-08-00710 May 12, 2017 Time: 16:29 # 1 ORIGINAL RESEARCH published: 16 May 2017 doi: 10.3389/fpsyg.2017.00710 Postural Communication of Emotion: Perception of Distinct Poses of Five Discrete Emotions Lukas D. Lopez, Peter J. Reschke, Jennifer M. Knothe and Eric A. Walle* Psychological Sciences, University of California, Merced, Merced, CA, USA Emotion can be communicated through multiple distinct modalities. However, an often-ignored channel of communication is posture. Recent research indicates that bodily posture plays an important role in the perception of emotion. However, research examining postural communication of emotion is limited by the variety of validated emotion poses and unknown cohesion of categorical and dimensional ratings. The present study addressed these limitations. Specifically, we examined individuals’ (1) categorization of emotion postures depicting 5 discrete emotions (joy, sadness, fear, anger, and disgust), (2) categorization of different poses depicting the same discrete emotion, and (3) ratings of valence and arousal for each emotion pose. Findings revealed that participants successfully categorized each posture as the target emotion, including Edited by: disgust. Moreover, participants accurately identified multiple distinct poses within each Petri Laukka, emotion category. In addition to the categorical responses, dimensional ratings of Stockholm University, Sweden valence and arousal revealed interesting overlap and distinctions between and within Reviewed by: Alessia Celeghin, emotion categories. These findings provide the first evidence of an identifiable posture University of Turin, Italy for disgust and instantiate the principle of equifinality of emotional communication Matteo Candidi, Sapienza University of Rome, Italy through the inclusion of distinct poses within emotion categories. Additionally, the *Correspondence: dimensional ratings corroborated the categorical data and provide further granularity for Eric A.
    [Show full text]