Literature DB >> 30505068

Multimodal Affective Analysis Using Hierarchical Attention Strategy with Word-Level Alignment.

Kangning Yang1, Shiyu Fu1, Yue Gu1, Shuhong Chen1, Xinyu Li1, Ivan Marsic1.   

Abstract

Multimodal affective computing, learning to recognize and interpret human affect and subjective information from multiple data sources, is still challenging because:(i) it is hard to extract informative features to represent human affects from heterogeneous inputs; (ii) current fusion strategies only fuse different modalities at abstract levels, ignoring time-dependent interactions between modalities. Addressing such issues, we introduce a hierarchical multimodal architecture with attention and word-level fusion to classify utterance-level sentiment and emotion from text and audio data. Our introduced model outperforms state-of-the-art approaches on published datasets, and we demonstrate that our model's synchronized attention over modalities offers visual interpretability.

Entities:  

Year:  2018        PMID: 30505068      PMCID: PMC6261375     

Source DB:  PubMed          Journal:  Proc Conf Assoc Comput Linguist Meet        ISSN: 0736-587X


  1 in total

1.  Combining Video, Audio and Lexical Indicators of Affect in Spontaneous Conversation via Particle Filtering.

Authors:  Arman Savran; Houwei Cao; Miraj Shah; Ani Nenkova; Ragini Verma
Journal:  Proc ACM Int Conf Multimodal Interact       Date:  2012
  1 in total
  7 in total

1.  Region Dual Attention-Based Video Emotion Recognition.

Authors:  Xiaodong Liu; Huating Xu; Miao Wang
Journal:  Comput Intell Neurosci       Date:  2022-06-15

2.  Human Conversation Analysis Using Attentive Multimodal Networks with Hierarchical Encoder-Decoder.

Authors:  Yue Gu; Xinyu Li; Kaixiang Huang; Shiyu Fu; Kangning Yang; Shuhong Chen; Moliang Zhou; Ivan Marsic
Journal:  Proc ACM Int Conf Multimed       Date:  2018-10

3.  Multimodal Attention Network for Trauma Activity Recognition from Spoken Language and Environmental Sound.

Authors:  Yue Gu; Ruiyu Zhang; Xinwei Zhao; Shuhong Chen; Jalal Abdulbaqi; Ivan Marsic; Megan Cheng; Randall S Burd
Journal:  IEEE Int Conf Healthc Inform       Date:  2019-11-21

4.  LGCCT: A Light Gated and Crossed Complementation Transformer for Multimodal Speech Emotion Recognition.

Authors:  Feng Liu; Si-Yuan Shen; Zi-Wang Fu; Han-Yang Wang; Ai-Min Zhou; Jia-Yin Qi
Journal:  Entropy (Basel)       Date:  2022-07-21       Impact factor: 2.738

5.  Research on Integration of Emotion Analysis in English Modular Teaching Based on Natural Language Processing.

Authors:  Fuxing Su
Journal:  Front Psychol       Date:  2022-07-22

6.  Words Can Shift: Dynamically Adjusting Word Representations Using Nonverbal Behaviors.

Authors:  Yansen Wang; Ying Shen; Zhun Liu; Paul Pu Liang; Amir Zadeh; Louis-Philippe Morency
Journal:  Proc Conf AAAI Artif Intell       Date:  2019-07

7.  Affective Latent Representation of Acoustic and Lexical Features for Emotion Recognition.

Authors:  Eesung Kim; Hyungchan Song; Jong Won Shin
Journal:  Sensors (Basel)       Date:  2020-05-04       Impact factor: 3.576

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.