HOW DEEP NEURAL NETWORKS CAN IMPROVE EMOTION RECOGNITION ON VIDEO DATA Pooya Khorrami1, Tom Le Paine1, Kevin Brady2, Charlie Dagli2, Thomas S. Huang1 1 Beckman Institute, University of Illinois at Urbana-Champaign 2 MIT Lincoln Laboratory 1 fpkhorra2, paine1, [email protected] 2 fkbrady, [email protected] ABSTRACT An alternative way to model the space of possible emo- There have been many impressive results obtained us- tions is to use a dimensional approach [11] where a person’s ing deep learning for emotion recognition tasks in the last emotions can be described using a low-dimensional signal few years. In this work, we present a system that per- (typically 2 or 3 dimensions). The most common dimensions forms emotion recognition on video data using both con- are (i) arousal and (ii) valence. Arousal measures how en- volutional neural networks (CNNs) and recurrent neural net- gaged or apathetic a subject appears while valence measures works (RNNs). We present our findings on videos from how positive or negative a subject appears. the Audio/Visual+Emotion Challenge (AV+EC2015). In our Dimensional approaches have two advantages over cat- experiments, we analyze the effects of several hyperparam- egorical approaches. The first being that dimensional ap- eters on overall performance while also achieving superior proaches can describe a larger set of emotions. Specifically, performance to the baseline and other competing methods. the arousal and valence scores define a two dimensional plane while the six basic emotions are represented as points in said Index Terms— Emotion Recognition, Convolutional plane. The second advantage is dimensional approaches can Neural Networks, Recurrent Neural Networks, Deep Learn- output time-continuous labels which allows for more realistic ing, Video Processing modeling of emotion over time. This could be particularly useful for representing video data. 1. INTRODUCTION Given the success of deep neural networks on datasets with categorical labels [12, 13, 10], one can ask the very nat- For several decades, emotion recognition has remained one ural question: is it possible to train a neural network to learn of the of the most important problems in the field of human a representation that is useful for dimensional emotion recog- computer interaction. A large portion of the community has nition in video data? focused on categorical models which try to group emotions In this paper, we will present two different frameworks into discrete categories. The most famous categories are the for training an emotion recognition system using deep neu- six basic emotions originally proposed by Ekman in [1, 2]: ral networks. The first is a single frame convolutional neural anger, disgust, fear, happiness, sadness, and surprise. These network (CNN) and the second is a combination of CNN and emotions were selected because they were all perceived simi- a recurrent neural network (RNN) where each input to the larly regardless of culture. RNN is the fully-connected features of a single frame CNN. Several datasets have been constructed to evaluate au- While many works have considered the benefits of using ei- tomatic emotion recognition systems such as the extended ther a CNN or a temporal model like an RNN or Long Short Cohn-Kanade (CK+) dataset [3] the MMI facial expression Term Memory (LSTM) network individually, very few works database [4] and the Toronto Face Dataset (TFD) [5]. In [14, 15] have considered the benefits of training with both the last few years, several methods based on hand-crafted types of networks combined. and, later, learned features [6, 7, 8, 9, 10] have performed quite well in recognizing the six basic emotions. Unfortu- Thus, the main contributions of this work are as follows: nately, these six basic emotions do not cover the full range of emotions that a person can express. 1. We train both a single-frame CNN and a CNN+RNN model and analyze their effectiveness on the dimen- The MIT Lincoln Laboratory part of this collaborative work is sponsored sional emotion recognition task. We also conduct ex- by the Assistant Secretary of Defense for Research & Engineering under Air Force Contract FA8721-05-C-0002. The Tesla K40 GPU used for this re- tensive analysis on the various hyperparameters of the search was donated by the NVIDIA Corporation. CNN+RNN model to support our chosen architecture. 2. We evaluate our models on the AV+EC 2015 dataset Input Conv. Layer 1 Conv. Layer 2 Conv. Layer 3 FC Regression [16] and demonstrate that our techniques can achieve comparable or superior performance to the baseline model and other competing methods. 96x96x1 64@5x5 128@5x5 256@5x5 2. DATASET Max Pooling Max Pooling Quadrant Pooling The AV+EC 2015 [16] corpus uses data from the RECOLA dataset [17], a multimodal corpus designed to monitor sub- Fig. 1. Single Frame CNN Architecture - Similar to net- jects as they worked in pairs remotely to complete a collab- work in [10], our network consists of three convolutional3 orative task. The type of modalities included: audio, video, layers containing 64, 128, and 256 filters, respectively, each electro-cardiogram (ECG) and electro-dermal activity (EDA). of size 5x5 followed by ReLU (Rectified Linear Unit) acti- These signals were recorded for 27 French-speaking subjects. vation functions. We add 2x2 max pooling layers after the The dataset contains two types of dimensional labels (arousal first two convolutional layers and quadrant pooling after the and valence) which were annotated by 6 people. Each dimen- third. The three convolutional layers are followed by a fully- sional label ranges from [−1; 1]. The dataset is partitioned connected layer containing 300 hidden units and a linear re- into three sets: train, development, and test, each containing gression layer. 9 subjects. In our experiments, we focus on predicting the valence score using just the video modality. Also, since the test set Output (t-W) Output (t-1) Output (t) labels were not readily available, we evaluate all of our ex- periments on the development set. We evaluate our tech- niques by computing three metrics: (i) Root Mean Square Error (RMSE) (ii) Pearson Correlation Coefficient (CC) and RNN RNN RNN (iii) Concordance Correlation Coefficient (CCC). The Con- cordance Correlation Coefficient tries to measure the agree- ment between two variables using the following expression: CNN CNN CNN 2ρσxσy ρc = 2 2 2 (1) σx + σy + (µx − µy) 2 2 where ρ is the Pearson correlation coefficient, σx and σy are the variance of the predicted and ground truth values re- IMG (t-W) IMG (t-1) IMG (t) spectively and µx and µy are their means, respectively. The strongest method is selected based on whichever obtains the highest CCC value. Fig. 2. CNN+RNN Network Architecture: Given a time t in a video, we extract a window of length W frames ([t − W; t]). 3. OUR APPROACH We model our single frame CNN as a feature extractor by fixing all of the parameters and removing the top regression 3.1. Single Frame Regression CNN layer. We then pass each frame within the window to the CNN and extract a 300 dimensional feature for every frame, each of The first model that we train is a single frame CNN. At each which is passed as an input to one node of the RNN. We then time point in a video, we pass the corresponding video frame take the valence score generated by the RNN at time t. through a CNN, shown visually in Figure 1. The CNN has 3 convolutional layers consisting of 64, 128, and 256 filters re- spectively, each of size 5x5. The first two convolutional layers did not use any form of annealing. All of our CNN models are followed by 2x2 max pooling while the third layer is fol- were trained using the anna software library 1. lowed by quadrant pooling. After the convolutional layers is a fully-connected layer with 300 hidden units and a linear re- gression layer to estimate the dimensional label. We use the 3.2. Adding Recurrent Neural Networks (RNNs) mean squared error (MSE) as our cost function. Despite having the ability to learn useful features directly All of the CNNs were trained using stochastic gradient from the video data, the single frame regression CNN com- descent with batch size of 128, momentum equal to 0.9, and weight decay of 1e-5. We used a constant learning of 0.01 and 1https://github.com/ifp-uiuc/anna 0.7 pletely ignores temporal information. Similar to the model in Ground Truth [15], we propose to incorporate the temporal information by CNN Prediction 0.6 CNN+RNN Prediction using a recurrent neural network (RNN) to propagate infor- mation from one time point to next. 0.5 We first model the CNN as a feature extractor by fixing 0.4 all of its parameters and removing the regression layer. Now, when a frame is passed to the CNN, we extract a 300 dimen- 0.3 sional vector from the fully-connected layer. For a given time Valence Score t, we take W frames from the past (i.e. [t − W; t]). We then 0.2 pass each frame from time t − W to t to the CNN and extract 0.1 W vectors in total, each length of 300. Each of these vectors is then passed as input to a node of the RNN. Each node in the 0.0 RNN then regresses the output label. We visualize the model 0.1 0 1000 2000 3000 4000 5000 6000 in Figure 2. Once again we use the mean squared error (MSE) Frame Number as our cost function during optimization. We train our RNN models with stochastic gradient de- Fig.
Details
-
File Typepdf
-
Upload Time-
-
Content LanguagesEnglish
-
Upload UserAnonymous/Not logged-in
-
File Pages5 Page
-
File Size-