<<

Multimodal human behavior analysis: Learning correlation and across modalities

The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters.

Citation Yale Song, Louis-Philippe Morency, and Randall Davis. 2012. Multimodal human behavior analysis: learning correlation and interaction across modalities. In Proceedings of the 14th ACM international conference on Multimodal interaction (ICMI '12). ACM, New York, NY, USA, 27-30.

As Published http://dx.doi.org/10.1145/2388676.2388684

Version Author's final manuscript

Citable link http://hdl.handle.net/1721.1/86099

Terms of Use Creative Commons Attribution-Noncommercial-Share Alike

Detailed Terms http://creativecommons.org/licenses/by-nc-sa/4.0/ Multimodal Human Behavior Analysis: Learning Correlation and Interaction Across Modalities

Yale Song Louis-Philippe Morency Randall Davis MIT CSAIL USC ICT MIT CSAIL Cambridge, MA 02139 Los Angeles, CA 90094 Cambridge, MA 02139 [email protected] [email protected] [email protected]

ABSTRACT learning the complex relationship across modalities is non-trivial. Multimodal human behavior analysis is a challenging task due to Figure 1(a) shows a pair of time-aligned sequences with audio and the presence of complex nonlinear correlations and interactions across visual features (from [11]; details can be found in Section 4.1). modalities. We present a novel approach to this problem based on When learning with this type of data, it is important to consider the Kernel Analysis (KCCA) and Multi-view correlation and interaction across modalities: An underlying cor- Hidden Conditional Random Fields (MV-HCRF). Our approach relation structure between modalities may make the amplitude of uses a nonlinear kernel to map multimodal data to a high-dimensional the signal from one modality different in relation to the signal from feature space and finds a new projection of the data that maximizes another modality, e.g., loud voice with exaggerated gestures. Also, the correlation across modalities. We use a multi-chain structured the interaction between modalities may have certain patterns that with disjoint sets of latent variables, one set per change the direction in which each sequence may evolve over time. modality, to jointly learn both view-shared and view-specific sub- In this paper, we investigate the hypothesis that transforming structures of the projected data, capturing interaction across modal- the original data to be maximally correlated across modalities, in ities explicitly. We evaluate our approach on a task of agreement a statistical sense, and capturing the interaction across modalities and disagreement recognition from nonverbal audio-visual cues us- explicitly from the transformed data improves recognition perfor- ing the Canal 9 dataset. Experimental results show that KCCA mance on human behavior analysis. To this end, we present a makes capturing nonlinear hidden dynamics easier and MV-HCRF novel approach to multimodal data learning that captures complex helps learning interaction across modalities. nonlinear correlations and interactions across modalities, based on KCCA [3] and MV-HCRF [10]. Our approach uses a nonlinear ker- nel to map multimodal data to a high-dimensional feature space and Categories and Subject Descriptors finds a new projection of the data that maximizes the correlation I.5.4 []: Applications—Signal processing across modalities. Figure 1(b) shows the projected signals found by KCCA, where the relative importance of gestures ‘head shake’ Keywords and ‘shoulder shrug’ have been emphasized to make the statistical relevance between the audio and visual signals become as clear as Multimodal signal processing; multi-view latent variable discrimi- possible. We then capture the interaction across modalities using native models; canonical correlation analysis; kernel methods a multi-chain structured latent variable discriminative model. The model uses disjoint sets of latent variables, one set per view, and 1. INTRODUCTION jointly learns both view-shared and view-specific sub-structures of Human communication is often accompanied by multimodal non- the projected data. verbal cues, such as gestures, eye gaze, and facial expressions. We evaluated our approach using the Canal 9 dataset [11], where These nonverbal cues play an important role in the way we com- the task is to recognize agreement and disagreement from nonver- municate with others and can convey as much information as spo- bal audio-visual cues. We report that using KCCA with MV-HCRF ken language. They complement or substitute for spoken language, to learn correlation and interaction across modalities successfully help to illustrate or emphasize main points, and provide a rich improves recognition performance compared to baseline methods. source of predictive information for understanding the intentions of the others. Automatic analysis of human behavior can thus ben- 2. RELATED WORK efit from harnessing multiple modalities. From the machine learning point of view, multimodal human be- Due to its theoretical and practical importance, multimodal hu- havior analysis continues to be a challenging task, in part because man behavior analysis has been a popular research topic. While audio-visual speech recognition is probably the most well known and successful example [6], multimodal affect recognition has re- cently been getting considerable attention [12]. Bousmalis et al. [1] Permission to make digital or hard copies of all or part of this work for proposed a system for spontaneous agreement and disagreement personal or classroom use is granted without fee provided that copies are recognition based only on prosodic and gesture cues, as we did not made or distributed for profit or commercial advantage and that copies here. They used an HCRF to capture the hidden dynamics of the bear this notice and the full citation on the first page. To copy otherwise, to multimodal cues. However, their approach did not consider the cor- republish, to post on servers or to redistribute to lists, requires prior specific relation and interaction across modalities explicitly. permission and/or a fee. ICMI’12, October 22Ð26, 2012, Santa Monica, California, USA. Canonical correlation analysis (CCA) has been successfully ap- Copyright 2012 ACM 978-1-4503-1467-1/12/10 ...$15.00. plied to multimedia content analysis [3, 13]. Hardoon et al. [3] 1 0.1 Audio Ԣܞ Ԣǡ܉݌ ݕൌሾƒ‰”‡‡ǡ†‹•ƒ‰”‡‡ሿȁ Features 0.5 0.05 ݕ

ܽ ܽ ܽ ݐ 0 ݐ 0 0 10 20 30 40 0 10 20 30 40 ܐଵ ܐଶ … ܐ୘

ݒܐ … ݒ ݒ 1 Visual KCCA 0.6 ܐଵ ܐଶ ୘ 0.4 ܉Ԣ ܉Ԣ ܉Ԣ Features 0.2 ଵ ଶ ୘ 0.5 0 −0.2 ܞԢଵ ܞԢଶ ܞԢ୘ ݐ −0.4 ݐ 0 0 10 20 30 40 0 10 20 30 40 (a) Original Data (b) Projected Data (c) MV-HCRF

Figure 1: An overview of our approach. (a) An audio-visual observation sequence from the Canal 9 dataset [11]. KCCA uses a nonlinear kernel to map the original data to a high-dimensional feature space, and finds a new projection of the data in the feature space that maximizes the correlation between audio and visual channels. (b) The projected data shows that emphasizing the amplitude of the ‘head shake’ and ‘shoulder shrug’ gestures maximized the correlation between audio and visual channels. (c) multi- view HCRF for jointly learning both view-shared and view-specific sub-structures of the projected data. at and vt are observation a v variables from audio and visual channels, and ht and ht are hidden variables for audio and visual channels. used kernel CCA (KCCA) for learning the semantic representation where ·, · denotes an inner product, and Φ is a nonlinear mapping of images and their associated text. However, their approach did function to a Hilbert space F, Φ:x ∈ R → Φ(x) ∈F. not consider capturing hidden dynamics in the data. Latent vari- To apply the kernel trick, the standard KCCA then rewrites wa able discriminative models, e.g., HCRF [7], have shown promising (and wv) as an inner product of the data A (and V) with a direction results in human behavior analysis tasks, for their ability to capture α (and β), the hidden dynamics (e.g., spatio-temporal dynamics). Recently, wa = A α, wv = V β the multi-view counterpart [10] showed a significant improvement (1) over single-view methods in recognizing human actions. However, If we assume that a’s and v’s are centered (i.e., zero), the their work did not learn nonlinear correlation across modalities. We goal is to maximize the correlation coefficient extend this body of work, enabling it to modeling multimodal hu- man behavior analysis. E[wa av wv] ρ(·, ·)= max  w ,w a v E[wa aa wa] E[wv vv wv] 3. OUR APPROACH wa AV wv N =max D { xi i }i=1 wa,wv Consider a labeled sequential dataset = ( ,y ) where wa AA wawv VV wv xi is a multivariate observation sequence and yi ∈Yis a sequence αAAVVβ label from a finite label set Y. Since we have audio-visual data, =max na×T α,β we use the notation xi =(ai, vi) where ai ∈ R and vi ∈ αAA AA α · βVV VV β n ×T R v T are audio and visual sequences of length , respectively. α KaKvβ Figure 1 shows an overview of our approach. We first find a new =max . (2) α,β 2 · 2 projection of the original observation sequence x =(a, v) using α Kaα β Kv β a v KCCA [3] such that the correlation between and is maximized where Ka = K(A, A) and Kv = K(V, V) are kernel matrices. in the projected space (Section 3.1). Then we use this projected Since Equation 2 is scale invariant with respect to α and β (they data as an input to MV-HCRF [10] to capture hidden dynamics and cancel out), the optimization problem is equivalent to: interaction between audio and visual data (Section 3.2). 2 2 max α KaKvβ subject to α Kaα = β Kv β =1 (3) 3.1 KCCA α,β N Given a set of paired samples {(ai, vi)}i=1, A =[a1, ··· , aN ] The corresponding Lagrangian dual form is and V =[v1, ··· , vN ], Canonical Correlation Analysis (CCA) aims to find a pair of transformations wa and wv such that the cor- θα 2 θβ 2 L(α, β, θ)=α KaKv β − (α Kaα − 1) − (β Kv β − 1) relation between the corresponding projections ρ(wa A, wv V) is 2 2 (4) maximized. However, since CCA finds wa and wv that are linear in the vector space, it may not reveal nonlinear relationships in the The solution to Equation 3 is found by taking derivatives of Equa- data [9]. tion 4 with respect to α and β, and solving a standard eigenvalue Kernel CCA (KCCA) [3] uses the kernel trick [9] to overcome problem [5]. However, when Ka and Kv are non-invertible, as is this limitation by projecting the original data onto a high dimen- common in practice with large datasets, problems can arise such sional feature space before running CCA. A kernel is a function as computational issues or degeneracy. This problem is solved by applying the partial Gram-Schmidt orthogonalization (PGSO) with K(xi, xj ) that, for all xi, xj ∈ R, a precision parameter η to reduce the dimensionality of the kernel K(xi, xj )=Φ(xi), Φ(xj ) matrices and approximate the correlation. After we find α and β, we plug the solution back in to Equation 1 Canal9 dataset is a collection of 72 political debates recorded by to obtain wa and wv, and finally obtain new projections: the Canal 9 TV station in Switzerland, with a total of roughly 42 hours of recordings. In each debate there is a moderator and A =[wa, a1 , ··· , wa, aN ] (5) two groups of participants who argue about a controversial politi- V =[wv, v1 , ··· , wv, vN ] . cal question. The dataset contains highly spontaneous verbal and 3.2 Multi-view HCRF non-verbal multimodal human behavior data. In order to facilitate comparison, we used the same part of the Given the new projection’s audio-visual features A and V (Equa- Canal9 dataset with nonverbal audio-visual features as was selected tion 5), the next step is to learn the hidden dynamics and interaction for use in Bousmalis et al. [1]. This consisted of 53 episodes of across modalities (see Figure 1 (b) and (c)). agreements and 94 episodes of disagreement collected over 11 de- A Multi-View Hidden [10] (MV-HCRF) bates. Bousmalis et al. selected the episodes based on two criteria: is a conditional that factorizes according to (a) the verbal content clearly indicates agreement/disagreement, a multi-chain structured undirected graph G =(V, E), where each which ensures that the ground truth label for the episode is known; chain is a discrete representation of each view. We use disjoint sets a a v v (b) the episode includes only one person, with a close-up shot of of latent variables h ∈H for audio and h ∈H for visual the speaker. The audio channel was encoded with 2-dimensional channel to learn both view-shared and view-specific sub-structures prosodic features, including the fundamental frequency (F0) and in audio-visual data. An MV-HCRF defines p(y | a , v ) as the energy. The visual channel was encoded with 8 gestures: ‘Head  exp{ΛΦ(y,h, a, v)} Nod’, ‘Head Shake’, ‘Forefinger Raise’, ‘Forefinger Raise-Like’, | a v  h p(y , )=  ‘Forefinger Wag’, ‘Hand Wag’, ‘Hands Scissor’, and ‘Shoulder  exp{Λ Φ(y , h, a , v )} y ,h Shrug’, where the presence/absence of the actions in each frame where h = {ha, hv} and Λ=[λ, ω] is a model parameter vector. was manually annotated with binary values. We downsampled the The function ΛΦ(y,h, a, v) is factorized with feature functions data from the original rate of 25 kHz to 12.5 kHz. fk(·) and gk(·) as    4.2 Methodology Λ Φ(y,h, a , v )= λkfk(y,hs, a , v ) The first step in our approach is to run KCCA to obtain a new s∈V k   projection of the data. We used a Gaussian RBF kernel as our 2 + ωkgk(y,hs,ht, a , v ). kernel function K(xi, xj )=exp −xi − xj  /2γ because of (s,t)∈E k its empirical success in the literature [9]. We validated the kernel k − Following [10], we define three types of feature functions. The la- width γ =10 ,k =[ 1, 0, 1] and the PGSO precision parameter η =[1:6]using grid search. The optimal parameter values were bel feature function fk(y,hs) encodes the relationship between a chosen based on the maximum correlation coefficient value. latent variable hs and a label y. The observation feature function Our followed a leave-two-debates-out cross-validation fk(hs, a , v ) encodes the relationship between a latent variable approach, where we selected 2 debates of the 11 debates as the test hs and observations x. The edge feature function gk(y,hs,ht) split, 3 debates for the validation split, and the remaining 6 de- encodes the transition between two latent variables hs and ht.We use the linked topology from [10] to define the edge set E (shown in bates for the training split. This was repeated five times on the 11 Figure 1(c)), which models contemporaneous connections between debates. The F1 scores were averaged to get the final result. We audio and visual observations, i.e., the concurrent latent states in chose four baseline models: Hidden Markov Models (HMM) [8], the audio and visual channel mutually affect each other. Note that Conditional Random Fields (CRF) [4], Hidden Conditional Ran- dom Fields (HCRF) [7], and Multi-view HCRF (MV-HCRF) [10]. the fk(·) are modeled under the assumption that views are condi- tionally independent given respective sets of latent variables, and We compared this to our KCCA with MV-HCRF approach. Note thus encode the view-specific sub-structures. The feature function that HMM and CRF perform per-frame classification, while HCRF and MV-HCRF perform per-sequence classification. The classifi- gk(·) encodes both view-shared and view-specific sub-structures. The optimal parameter set Λ∗ is found by minimizing the nega- cation results of each model in turn were measured accordingly. tive conditional log-likelihood We automatically validated the hyper-parameters of all models. For all CRF-family models, we varied the L2-norm regularization N k 1 2 factor σ =10,k =[0, 1, 2] (see Equation 6). For HMM and min L(Λ) = Λ − log p(yi | ai, vi;Λ) (6) |H| Λ 2σ2 HCRF, the number of hidden states were varied =[2:6]; for i=1 MV-HCRF, they were |HA|, |HV | =([2:4], [2 : 4]). Since the where the first term is the Gaussian prior over Λ that works as an optimization problems in HMM, HCRF and MV-HCRF are non- ∗ L2-norm regularization. We find the optimal parameters Λ us- convex, we performed two random initializations of each model; ing gradient descent with a quasi-newton optimization method, the the best model was selected based on the F1 score on the validation limited-memory BFGS algorithm [5]. The marginal probability of split. The L-BFGS solver was set to terminate after 500 iterations. each node is obtained by performing an inference task using the Junction Tree algorithm [2]. 4.3 Result and Discussion We first compared our approach to existing methods: HMM [8], 4. CRF [4], HCRF [7], and MV-HCRF [10]. Figure 2 shows a bar plot In this section, we describe the dataset, detail our experimental of mean F1 scores and their standard deviations. We can see that methodology, and discuss our results. our approach clearly outperforms all other models. For further analysis, we investigated whether learning nonlinear 4.1 Dataset correlation was important, comparing KCCA to CCA and the origi- We evaluated our approach using the Canal9 dataset [11], where nal data. Table 1 shows that models trained with KCCA always out- the task is to recognize agreement and disagreement from nonver- performed the others, suggesting that learning nonlinear correlation bal audio-visual cues during spontaneous political debates. The in the data was important. Figure 1(b) shows the data projected in Models Audio Video Audio+Video

0.8 HMM .54 (.08) .58 (.11) .59 (.09) CRF .48 (.05) .58 (.15) .61 (.04) 0.7 HCRF .52 (.09) .60 (.09) .64 (.13) MV-HCRF · · .68 (.13) 0.6 KCCA + MV-HCRF · · .72 (.07) Mean F1 Score 0.5 .59 .61 .64 .68 .72 Table 2: Experimental results ( and standard deviations HMM CRF HCRF MV−HCRF Our Approach of F1 scores) comparing unimodal (audio or video) features to the audio-visual features. The results confirms that using both Figure 2: A bar plot of mean F1 scores with error bars showing audio and visual features are important in our task. standard deviations. This shows empirically that our approach successfully learned correlations and interactions between au- dio and visual features using KCCA and MV-HCRF. 6. ACKNOWLEDGMENTS This work was funded by ONR grant #N000140910625, NSF grant #IIS-1118018, NSF grant #IIS-1018055, and U.S. Army RDE- Models Original Data CCA KCCA COM. HMM .59 (.09) .59 (.12) .61 (.13) CRF .61 (.04) .63 (.03) .67 (.08) 7. REFERENCES HCRF .64 (.13) .65 (.07) .69 (.06) [1] K. Bousmalis, L.-P. Morency, and M. Pantic. Modeling MV-HCRF .68 (.13) .71 (.07) .72 (.07) hidden dynamics of multimodal cues for spontaneous agreement and disagreement recognition. In FG, 2011. [2] R. G. Cowell, A. P. Dawid, S. L. Lauritzen, and D. J. Table 1: Experimental results (means and standard deviations Spiegelhalter. Probabilistic Networks and Expert Systems. of F1 scores) comparing KCCA to CCA and the original data. Springer-Verlag, 1999. The results show that learning nonlinear correlation in the data [3] D. R. Hardoon, S. Szedmák, and J. Shawe-Taylor. Canonical was important in our task. correlation analysis: An overview with application to learning methods. Neural Comp., 16(12):2639Ð2664, 2004. a new space found by KCCA, where the ‘head shake’ and ‘shoul- [4] J. D. Lafferty, A. McCallum, and F. C. N. Pereira. der shrug’ gestures were relatively emphasized compared to ‘head Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In ICML, 2001. nod’, which maximized the correlation between the audio and vi- sual signals. We believe that this made our data more descriptive, [5] J. Nocedal and S. J. Wright. Numerical Optimization. allowing the learning algorithm to capture the hidden dynamics and Springer-Verlag, 1999. interactions between modalities more effectively. [6] G. Potamianos, C. Neti, J. Luettin, and I. Matthews. We also investigated whether our approach captures interaction Audio-Visual Automatic Speech Recognition: An Overview. between audio-visual signals successfully. We compared the mod- MIT Press, 2004. els trained with a unimodal feature (audio or visual) to the models [7] A. Quattoni, S. B. Wang, L.-P. Morency, M. Collins, and trained with audio-visual features. Table 2 shows means and stan- T. Darrell. Hidden conditional random fields. TPAMI, dard deviations of the F1 scores. In the three single-chain mod- 29(10):1848Ð1852, 2007. els, HMM, CRF, and HCRF, there was an improvement when both [8] L. R. Rabiner. A tutorial on hidden Markov models and audio and visual features were used, confirming that using a com- selected applications in speech recognition, pages 267Ð296. bination of audio and visual features for our task is indeed impor- Morgan Kaufmann Publishers Inc., 1990. tant. Also, MV-HCRF outperformed HMM, CRF, HCRF, showing [9] B. Scholkopf and A. J. Smola. Learning with Kernels: empirically that learning interaction between audio-visual signals Support Vector Machines, Regularization, Optimization, and explicitly improved the performance. Beyond. MIT Press, Cambridge, MA, USA, 2001. [10] Y. Song, L.-P. Morency, and R. Davis. Multi-view latent 5. CONCLUSIONS variable discriminative models for action recognition. In CVPR, 2012. We presented a novel approach to multimodal human behavior [11] A. Vinciarelli, A. Dielmann, S. Favre, and H. Salamin. analysis using KCCA and MV-HCRF, and evaluated it on a task Canal9: A database of political debates for analysis of social of recognizing agreement and disagreement from nonverbal audio- interactions. In ACII, 2009. visual cues using the Canal9 dataset. On this dataset, we showed [12] Z. Zeng, M. Pantic, G. I. Roisman, and T. S. Huang. A that preprocessing multimodal data with KCCA, by projecting the survey of affect recognition methods: Audio, visual, and data to a new space where the correlation across modalities is max- spontaneous expressions. TPAMI, 31(1):39Ð58, 2009. imized, helps capture complex nonlinear relationship in the data. We also showed that KCCA with MV-HCRF, which jointly learns [13] H. Zhang, Y. Zhuang, and F. Wu. Cross-modal correlation view-shared and view-specific interactions explicitly, improves the learning for clustering on image-audio dataset. In MM, pages recognition performance, showing that our approach successfully 273Ð276, 2007. captured complex nonlinear interaction across modalities. We look forward to applying our technique to other applications in multi- modal human behavior analysis for further analysis.