Literature DB >> 25300451

Combining Video, Audio and Lexical Indicators of Affect in Spontaneous Conversation via Particle Filtering.

Arman Savran1, Houwei Cao2, Miraj Shah, Ani Nenkova, Ragini Verma.   

Abstract

We present experiments on fusing facial video, audio and lexical indicators for affect estimation during dyadic conversations. We use temporal statistics of texture descriptors extracted from facial video, a combination of various acoustic features, and lexical features to create regression based affect estimators for each modality. The single modality regressors are then combined using particle filtering, by treating these independent regression outputs as measurements of the affect states in a Bayesian filtering framework, where previous observations provide prediction about the current state by means of learned affect dynamics. Tested on the Audio-visual Emotion Recognition Challenge dataset, our single modality estimators achieve substantially higher scores than the official baseline method for every dimension of affect. Our filtering-based multi-modality fusion achieves correlation performance of 0.344 (baseline: 0.136) and 0.280 (baseline: 0.096) for the fully continuous and word level sub challenges, respectively.

Entities:  

Keywords:  adaboost; affective computing; class-spectral features; emotion dynamics; emotion recognition; lexical analysis; local binary patterns; multi-modality fusion; particle filtering; svm

Year:  2012        PMID: 25300451      PMCID: PMC4187218          DOI: 10.1145/2388676.2388781

Source DB:  PubMed          Journal:  Proc ACM Int Conf Multimodal Interact


  2 in total

Review 1.  A survey of affect recognition methods: audio, visual, and spontaneous expressions.

Authors:  Zhihong Zeng; Maja Pantic; Glenn I Roisman; Thomas S Huang
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2009-01       Impact factor: 6.226

2.  Class-Level Spectral Features for Emotion Recognition.

Authors:  Dmitri Bitouk; Ragini Verma; Ani Nenkova
Journal:  Speech Commun       Date:  2010-07       Impact factor: 2.017

  2 in total
  3 in total

1.  Multimodal Affective Analysis Using Hierarchical Attention Strategy with Word-Level Alignment.

Authors:  Kangning Yang; Shiyu Fu; Yue Gu; Shuhong Chen; Xinyu Li; Ivan Marsic
Journal:  Proc Conf Assoc Comput Linguist Meet       Date:  2018-07

2.  Acoustic and Lexical Representations for Affect Prediction in Spontaneous Conversations.

Authors:  Houwei Cao; Arman Savran; Ragini Verma; Ani Nenkova
Journal:  Comput Speech Lang       Date:  2015-01-01       Impact factor: 1.899

3.  Human Conversation Analysis Using Attentive Multimodal Networks with Hierarchical Encoder-Decoder.

Authors:  Yue Gu; Xinyu Li; Kaixiang Huang; Shiyu Fu; Kangning Yang; Shuhong Chen; Moliang Zhou; Ivan Marsic
Journal:  Proc ACM Int Conf Multimed       Date:  2018-10
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.