Literature DB >> 35983132

Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks.

Zhibang Quan1, Tao Sun1, Mengli Su1, Jishu Wei1.   

Abstract

Multimodal sentiment analysis has been an active subfield in natural language processing. This makes multimodal sentiment tasks challenging due to the use of different sources for predicting a speaker's sentiment. Previous research has focused on extracting single contextual information within a modality and trying different modality fusion stages to improve prediction accuracy. However, a factor that may lead to poor model performance is that this does not consider the variability between modalities. Furthermore, existing fusion methods tend to extract the representational information of individual modalities before fusion. This ignores the critical role of intermodal interaction information for model prediction. This paper proposes a multimodal sentiment analysis method based on cross-modal attention and gated cyclic hierarchical fusion network MGHF. MGHF is based on the idea of distribution matching, which enables modalities to obtain representational information with a synergistic effect on the overall sentiment orientation in the temporal interaction phase. After that, we designed a gated cyclic hierarchical fusion network that takes text-based acoustic representation, text-based visual representation, and text representation as inputs and eliminates redundant information through a gating mechanism to achieve effective multimodal representation interaction fusion. Our extensive experiments on two publicly available and popular multimodal datasets show that MGHF has significant advantages over previous complex and robust baselines.
Copyright © 2022 Zhibang Quan et al.

Entities:  

Mesh:

Year:  2022        PMID: 35983132      PMCID: PMC9381258          DOI: 10.1155/2022/4767437

Source DB:  PubMed          Journal:  Comput Intell Neurosci


  6 in total

1.  Integrating Multimodal Information in Large Pretrained Transformers.

Authors:  Wasifur Rahman; Md Kamrul Hasan; Sangwu Lee; Amir Zadeh; Chengfeng Mao; Louis-Philippe Morency; Ehsan Hoque
Journal:  Proc Conf Assoc Comput Linguist Meet       Date:  2020-07

2.  Long short-term memory.

Authors:  S Hochreiter; J Schmidhuber
Journal:  Neural Comput       Date:  1997-11-15       Impact factor: 2.026

3.  Multimodal Machine Learning: A Survey and Taxonomy.

Authors:  Tadas Baltrusaitis; Chaitanya Ahuja; Louis-Philippe Morency
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2018-01-25       Impact factor: 6.226

4.  Multimodal Transformer for Unaligned Multimodal Language Sequences.

Authors:  Yao-Hung Hubert Tsai; Shaojie Bai; Paul Pu Liang; J Zico Kolter; Louis-Philippe Morency; Ruslan Salakhutdinov
Journal:  Proc Conf Assoc Comput Linguist Meet       Date:  2019-07

5.  Multi-attention Recurrent Network for Human Communication Comprehension.

Authors:  Amir Zadeh; Paul Pu Liang; Soujanya Poria; Prateek Vij; Erik Cambria; Louis-Philippe Morency
Journal:  Proc Conf AAAI Artif Intell       Date:  2018-02

6.  Words Can Shift: Dynamically Adjusting Word Representations Using Nonverbal Behaviors.

Authors:  Yansen Wang; Ying Shen; Zhun Liu; Paul Pu Liang; Amir Zadeh; Louis-Philippe Morency
Journal:  Proc Conf AAAI Artif Intell       Date:  2019-07
  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.