A Review of Alexithymia and Emotion Perception in Music, Odor, Taste, and Touch
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Effect of Emotional Arousal 1 Running Head
Effect of Emotional Arousal 1 Running Head: THE EFFECT OF EMOTIONAL AROUSAL ON RECALL The Effect of Emotional Arousal and Valence on Memory Recall 500181765 Bangor University Group 14, Thursday Afternoon Effect of Emotional Arousal 2 Abstract This study examined the effect of emotion on memory when recalling positive, negative and neutral events. Four hundred and fourteen participants aged over 18 years were asked to read stories that differed in emotional arousal and valence, and then performed a spatial distraction task before they were asked to recall the details of the stories. Afterwards, participants rated the stories on how emotional they found them, from ‘Very Negative’ to ‘Very Positive’. It was found that the emotional stories were remembered significantly better than the neutral story; however there was no significant difference in recall when a negative mood was induced versus a positive mood. Therefore this research suggests that emotional valence does not affect recall but emotional arousal affects recall to a large extent. Effect of Emotional Arousal 3 Emotional arousal has often been found to influence an individual’s recall of past events. It has been documented that highly emotional autobiographical memories tend to be remembered in better detail than neutral events in a person’s life. Structures involved in memory and emotions, the hippocampus and amygdala respectively, are joined in the limbic system within the brain. Therefore, it would seem true that emotions and memory are linked. Many studies have investigated this topic, finding that emotional arousal increases recall. For instance, Kensinger and Corkin (2003) found that individuals remember emotionally arousing words (such as swear words) more than they remember neutral words. -
How Emotions Influence Consumers' Perception of Credibility and Trust in CSR Communication
Department of Informatics and Media Digital Media and Society Two-Year Master’s thesis The Age of Emotionality? – How emotions influence consumers’ perception of credibility and trust in CSR communication Student: Anika Reupsch Supervisor: Göran Svensson Spring 2017 Abstract Companies around the world are using different strategies for their corporate social responsibility (CSR) communication, but finding an appropriate strategy to enhance trust and credibility on the consumer side remains challenging. The constitutive aspect of emotions in CSR communication has long been overlooked. Therefore, this study investigates the influence emotions in CSR communication have on the credibility and trust consumers have in a firm’s CSR. Quantitative research with group division was conducted. The online survey used authentic CSR communication content of a German trading company, which was classified in two categories: low emotional appeal and high emotional appeal. Statistical tests for group differences, correlation, factor analysis and multiple regression were carried out to study the influence of emotions on credibility and trust. The applied test statistics for comparing groups differences have shown a significant difference for guilt and regret, which are reduced for the group exposed to high emotional CSR content. Similarly, comprehension is enhanced, whereas sincerity is reduced for high emotional content. Above that, it was proved that positive emotions have a positive effect on perceived credibility for both text- image and moving image CSR content. However, negative emotions influence perceived credibility only for the text-image content. To the contrary, the analysis for moving image content revealed another dimension formed by interest and compassion, that positively influences perceived credibility. -
Neuroticism and Facial Emotion Recognition in Healthy Adults
First Impact Factor released in June 2010 bs_bs_banner and now listed in MEDLINE! Early Intervention in Psychiatry 2015; ••: ••–•• doi:10.1111/eip.12212 Brief Report Neuroticism and facial emotion recognition in healthy adults Sanja Andric,1 Nadja P. Maric,1,2 Goran Knezevic,3 Marina Mihaljevic,2 Tijana Mirjanic,4 Eva Velthorst5,6 and Jim van Os7,8 Abstract 1School of Medicine, University of Results: A significant negative corre- Belgrade, 2Clinic for Psychiatry, Clinical Aim: The aim of the present study lation between the degree of neuroti- 3 Centre of Serbia, Faculty of Philosophy, was to examine whether healthy cism and the percentage of correct Department of Psychology, University of individuals with higher levels of neu- answers on DFAR was found only for Belgrade, Belgrade, 4Special Hospital for roticism, a robust independent pre- happy facial expression (significant Psychiatric Disorders Kovin, Kovin, Serbia; after applying Bonferroni correction). 5Department of Psychiatry, Academic dictor of psychopathology, exhibit Medical Centre University of Amsterdam, altered facial emotion recognition Amsterdam, 6Departments of Psychiatry performance. Conclusions: Altered sensitivity to the and Preventive Medicine, Icahn School of emotional context represents a useful Medicine, New York, USA; 7South Methods: Facial emotion recognition and easy way to obtain cognitive phe- Limburg Mental Health Research and accuracy was investigated in 104 notype that correlates strongly with Teaching Network, EURON, Maastricht healthy adults using the Degraded -
How Deep Neural Networks Can Improve Emotion Recognition on Video Data
HOW DEEP NEURAL NETWORKS CAN IMPROVE EMOTION RECOGNITION ON VIDEO DATA Pooya Khorrami1, Tom Le Paine1, Kevin Brady2, Charlie Dagli2, Thomas S. Huang1 1 Beckman Institute, University of Illinois at Urbana-Champaign 2 MIT Lincoln Laboratory 1 fpkhorra2, paine1, [email protected] 2 fkbrady, [email protected] ABSTRACT An alternative way to model the space of possible emo- There have been many impressive results obtained us- tions is to use a dimensional approach [11] where a person’s ing deep learning for emotion recognition tasks in the last emotions can be described using a low-dimensional signal few years. In this work, we present a system that per- (typically 2 or 3 dimensions). The most common dimensions forms emotion recognition on video data using both con- are (i) arousal and (ii) valence. Arousal measures how en- volutional neural networks (CNNs) and recurrent neural net- gaged or apathetic a subject appears while valence measures works (RNNs). We present our findings on videos from how positive or negative a subject appears. the Audio/Visual+Emotion Challenge (AV+EC2015). In our Dimensional approaches have two advantages over cat- experiments, we analyze the effects of several hyperparam- egorical approaches. The first being that dimensional ap- eters on overall performance while also achieving superior proaches can describe a larger set of emotions. Specifically, performance to the baseline and other competing methods. the arousal and valence scores define a two dimensional plane while the six basic emotions are represented as points in said Index Terms— Emotion Recognition, Convolutional plane. The second advantage is dimensional approaches can Neural Networks, Recurrent Neural Networks, Deep Learn- output time-continuous labels which allows for more realistic ing, Video Processing modeling of emotion over time. -
The Emotional Stroop Effect in Anxiety Disorders General Emotionality Or Disorder Specificity?
Anxiety Disorders 15 Q2001) 147±159 The emotional Stroop effect in anxiety disorders General emotionality or disorder specificity? Eni S. Beckera,*, Mike Rincka, JuÈrgen Margraf a, Walton T. Rothb aTU, Dresden, Germany bVA Palo Alto Health Care System, Palo Alto, CA, USA Received 8 January 1999; received in revised form 22 April 1999; accepted 15 October 1999 Abstract Selective attentional biases, often documented with a modified Stroop task, are considered to play an important role in the etiology and maintenance of anxiety. Two competing explanations for these effects are selectivity for highly emotional words in general vs. selectivity for disorder-specific words. We tested these explanations in 32 patients with generalized anxiety disorder QGAD), 29 patients with social phobia QSP), and 31 non-anxious controls. Stimuli were of four kinds: GAD-related words, SP-related words, words with a neutral valence, and words with a positive valence. Different attentional biases were observed: GAD patients were slowed by all types of emotional words, while SP patients were distracted specifically by speech-related words. D 2001 Elsevier Science Inc. All rights reserved. Keywords: Attention; Generalized anxiety disorder; Social phobia; Stroop task 1. Introduction Cognitive models of anxiety have received increasing empirical support over the past decade Qe.g., Barlow 1988; Beck, Emery, & Greenberg, 1985; 1992). * Corresponding author. Department of Clinical Psychology, Dresden University of Technology, D-01062 Dresden, Germany. E-mail address: [email protected] QE.S. Becker). 0887-6185/01/$ ± see front matter D 2001 Elsevier Science Inc. All rights reserved. PII: S0887-6185Q01)00055-X 148 E.S. -
Emotion Recognition Depends on Subjective Emotional Experience and Not on Facial
Emotion recognition depends on subjective emotional experience and not on facial expressivity: evidence from traumatic brain injury Travis Wearne1, Katherine Osborne-Crowley1, Hannah Rosenberg1, Marie Dethier2 & Skye McDonald1 1 School of Psychology, University of New South Wales, Sydney, NSW, Australia 2 Department of Psychology: Cognition and Behavior, University of Liege, Liege, Belgium Correspondence: Dr Travis A Wearne, School of Psychology, University of New South Wales, NSW, Australia, 2052 Email: [email protected] Number: (+612) 9385 3310 Abstract Background: Recognising how others feel is paramount to social situations and commonly disrupted following traumatic brain injury (TBI). This study tested whether problems identifying emotion in others following TBI is related to problems expressing or feeling emotion in oneself, as theoretical models place emotion perception in the context of accurate encoding and/or shared emotional experiences. Methods: Individuals with TBI (n = 27; 20 males) and controls (n = 28; 16 males) were tested on an emotion recognition task, and asked to adopt facial expressions and relay emotional memories according to the presentation of stimuli (word & photos). After each trial, participants were asked to self-report their feelings of happiness, anger and sadness. Judges that were blind to the presentation of stimuli assessed emotional facial expressivity. Results: Emotional experience was a unique predictor of affect recognition across all emotions while facial expressivity did not contribute to any of the regression models. Furthermore, difficulties in recognising emotion for individuals with TBI were no longer evident after cognitive ability and experience of emotion were entered into the analyses. Conclusions: Emotion perceptual difficulties following TBI may stem from an inability to experience affective states and may tie in with alexythymia in clinical conditions. -
Improving Emotion Perception and Emotion Regulation Through a Web-Based Emotional Intelligence Training (WEIT) Program for Future Leaders
Volume 11, Number 2, November 2019 pp 17 - 32 www.um.edu.mt/ijee Improving Emotion Perception and Emotion Regulation Through a Web-Based Emotional Intelligence Training (WEIT) Program for Future Leaders 1 Christina Köppe, Marco Jürgen Held and Astrid Schütz University of Bamberg, Bamberg, Germany We evaluated a Web-Based Emotional Intelligence Training (WEIT) program that was based on the four-branch model of emotional intelligence (EI) and which aimed at improving emotion perception (EP) and emotion regulation (ER) in future leaders. Using a controlled experimental design, we evaluated the short-term (directly after the WEIT program) and long-term (6 weeks later) effects in a sample of 134 (59 training group [TG], 75 wait list control group [CG]) business students, and additionally tested whether WEIT helped to reduce perceived stress. For EP, WEIT led to a significant increase in the TG directly after training (whereas the wait list CG showed no change). Changes remained stable after 6 weeks in the TG, but there were no significant differences between the TG and CG at follow-up. By contrast, ER did not show an increase directly after WEIT, but 6 weeks later, the TG had larger improvements than the CG. The results mostly confirmed that emotional abilities can be increased through web-based training. Participants’ perceived stress did not decrease after the training program. Further refinement and validation of WEIT is needed. Keywords: emotional intelligence, web-based training, stress, future leaders First submission 15th May 2019; Accepted for publication 4th September 2019. Introduction Emotional intelligence (EI) has attracted considerable attention in recent years (see Côté, 2014). -
Facial Emotion Recognition
Issue 1 | 2021 Facial Emotion Recognition Facial Emotion Recognition (FER) is the technology that analyses facial expressions from both static images and videos in order to reveal information on one’s emotional state. The complexity of facial expressions, the potential use of the technology in any context, and the involvement of new technologies such as artificial intelligence raise significant privacy risks. I. What is Facial Emotion FER analysis comprises three steps: a) face de- Recognition? tection, b) facial expression detection, c) ex- pression classification to an emotional state Facial Emotion Recognition is a technology used (Figure 1). Emotion detection is based on the ana- for analysing sentiments by different sources, lysis of facial landmark positions (e.g. end of nose, such as pictures and videos. It belongs to the eyebrows). Furthermore, in videos, changes in those family of technologies often referred to as ‘affective positions are also analysed, in order to identify con- computing’, a multidisciplinary field of research on tractions in a group of facial muscles (Ko 2018). De- computer’s capabilities to recognise and interpret pending on the algorithm, facial expressions can be human emotions and affective states and it often classified to basic emotions (e.g. anger, disgust, fear, builds on Artificial Intelligence technologies. joy, sadness, and surprise) or compound emotions Facial expressions are forms of non-verbal (e.g. happily sad, happily surprised, happily disgus- communication, providing hints for human emo- ted, sadly fearful, sadly angry, sadly surprised) (Du tions. For decades, decoding such emotion expres- et al. 2014). In other cases, facial expressions could sions has been a research interest in the field of psy- be linked to physiological or mental state of mind chology (Ekman and Friesen 2003; Lang et al. -
Developing a Brain Specialized for Face Perception: a Converging
Michelle de Haan Institute of Child Health University College London Developing a Brain Developmental Cognitive Neurosciences Unit The Wolfson Centre Specialized for Face Mecklenburgh Square London WCIN ZAP United Kingdom Perception: A Converging Kate Humphreys MarkH. Johnson Methods Approach Centre for Brain & Cognitive Development, Birkbeck College 32 Torrington Square London WCIE 7JL United Kingdom Received 21 February 2000; Accepted 13 November 2001 ABSTRACT: Adults are normally very quick and accurate at recognizing facial identity.This skill has been explained by two opposing views as being due to the existence of an innate cortical ``face module'' or to the natural consequence of adults' extensive experience with faces.Neither of these views puts particular importance on studying development of face-processing skills, as in one view the module simply comes on-line and in the other view development is equated with adult learning. In this article, we present evidence from a variety of methodologies to argue for an ``interactive specialization'' view.In this view, orienting tendencies present early in life ensure faces are a frequent visual input to higher cortical areas in the ventral visual pathway.In this way, cortical specialization for face processing is an emergent product of the interaction of factors both intrinsic and extrinsic to the developing child. ß 2002 Wiley Periodicals, Inc. Dev Psychobiol 40: 200± 212, 2002. DOI 10.1002/dev.10027 Keywords: face processing; infancy; neuroimaging; cortical specialization; prosopagnosia INTRODUCTION investigators have proposed that there is an innate ``social brain'' with pathways and circuits genetically Our sophisticated ability to perceive and analyze in- prespeci®ed for processing social information -e.g., formation from the faces of our fellow humans under- Baron-Cohen et al., 1999). -
Alexithymia – Not Autism – Is Associated with Frequency of Social T Interactions in Adults ∗ Alan H
Behaviour Research and Therapy 123 (2019) 103477 Contents lists available at ScienceDirect Behaviour Research and Therapy journal homepage: www.elsevier.com/locate/brat Alexithymia – Not autism – is associated with frequency of social T interactions in adults ∗ Alan H. Gerbera, Jeffrey M. Girardb, Stacey B. Scotta, Matthew D. Lernera, a Department of Psychology, Stony Brook University, Stony Brook, NY, USA b Language Technologies Institute, Carnegie Mellon University, Pittsburgh, PA, USA ARTICLE INFO ABSTRACT Keywords: Objective: While much is known about the quality of social behavior among neurotypical individuals and those Alexithymia with autism spectrum disorder (ASD), little work has evaluated quantity of social interactions. This study used Autism ecological momentary assessment (EMA) to quantify in vivo daily patterns of social interaction in adults as a ASD function of demographic and clinical factors. EMA Method: Adults with and without ASD (NASD = 23, NNeurotypical = 52) were trained in an EMA protocol to report Social interaction their social interactions via smartphone over one week. Participants completed measures of IQ, ASD symptom severity and alexithymia symptom severity. Results: Cyclical multilevel models were used to account for nesting of observations. Results suggest a daily cyclical pattern of social interaction that was robust to ASD and alexithymia symptoms. Adults with ASD did not have fewer social interactions than neurotypical peers; however, severity of alexithymia symptoms predicted fewer social interactions -
Emotion Perception in Habitual Players of Action Video Games Swann Pichon, Benoit Bediou, Lia Antico, Rachael Jack, Oliver Garrod, Chris Sims, C
Emotion Emotion Perception in Habitual Players of Action Video Games Swann Pichon, Benoit Bediou, Lia Antico, Rachael Jack, Oliver Garrod, Chris Sims, C. Shawn Green, Philippe Schyns, and Daphne Bavelier Online First Publication, July 6, 2020. http://dx.doi.org/10.1037/emo0000740 CITATION Pichon, S., Bediou, B., Antico, L., Jack, R., Garrod, O., Sims, C., Green, C. S., Schyns, P., & Bavelier, D. (2020, July 6). Emotion Perception in Habitual Players of Action Video Games. Emotion. Advance online publication. http://dx.doi.org/10.1037/emo0000740 Emotion © 2020 American Psychological Association 2020, Vol. 2, No. 999, 000 ISSN: 1528-3542 http://dx.doi.org/10.1037/emo0000740 Emotion Perception in Habitual Players of Action Video Games Swann Pichon, Benoit Bediou, and Lia Antico Rachael Jack and Oliver Garrod University of Geneva, Campus Biotech Glasgow University Chris Sims C. Shawn Green Rensselaer Polytechnic Institute University of Wisconsin–Madison Philippe Schyns Daphne Bavelier Glasgow University University of Geneva, Campus Biotech Action video game players (AVGPs) display superior performance in various aspects of cognition, especially in perception and top-down attention. The existing literature has examined these performance almost exclusively with stimuli and tasks devoid of any emotional content. Thus, whether the superior performance documented in the cognitive domain extend to the emotional domain remains unknown. We present 2 cross-sectional studies contrasting AVGPs and nonvideo game players (NVGPs) in their ability to perceive facial emotions. Under an enhanced perception account, AVGPs should outperform NVGPs when processing facial emotion. Yet, alternative accounts exist. For instance, under some social accounts, exposure to action video games, which often contain violence, may lower sensitivity for empathy-related expressions such as sadness, happiness, and pain while increasing sensitivity to aggression signals. -
Deep-Learning-Based Models for Pain Recognition: a Systematic Review
applied sciences Article Deep-Learning-Based Models for Pain Recognition: A Systematic Review Rasha M. Al-Eidan * , Hend Al-Khalifa and AbdulMalik Al-Salman College of Computer and Information Sciences, King Saud University, Riyadh 11451, Saudi Arabia; [email protected] (H.A.-K.); [email protected] (A.A.-S.) * Correspondence: [email protected] Received: 31 July 2020; Accepted: 27 August 2020; Published: 29 August 2020 Abstract: Traditional standards employed for pain assessment have many limitations. One such limitation is reliability linked to inter-observer variability. Therefore, there have been many approaches to automate the task of pain recognition. Recently, deep-learning methods have appeared to solve many challenges such as feature selection and cases with a small number of data sets. This study provides a systematic review of pain-recognition systems that are based on deep-learning models for the last two years. Furthermore, it presents the major deep-learning methods used in the review papers. Finally, it provides a discussion of the challenges and open issues. Keywords: pain assessment; pain recognition; deep learning; neural network; review; dataset 1. Introduction Deep learning is an important branch of machine learning used to solve many applications. It is a powerful way of achieving supervised learning. The history of deep learning has undergone different naming conventions and developments [1]. It was first referred to as cybernetics in the 1940s–1960s. Then, in the 1980s–1990s, it became known as connectionism. Finally, the phrase deep learning began to be utilized in 2006. There are also some names that are based on human knowledge.