Literature DB >> 32257595

Multi-attention Recurrent Network for Human Communication Comprehension.

Amir Zadeh1, Paul Pu Liang1, Soujanya Poria2, Prateek Vij2, Erik Cambria2, Louis-Philippe Morency1.   

Abstract

Human face-to-face communication is a complex multimodal signal. We use words (language modality), gestures (vision modality) and changes in tone (acoustic modality) to convey our intentions. Humans easily process and understand face-to-face communication, however, comprehending this form of communication remains a significant challenge for Artificial Intelligence (AI). AI must understand each modality and the interactions between them that shape the communication. In this paper, we present a novel neural architecture for understanding human communication called the Multi-attention Recurrent Network (MARN). The main strength of our model comes from discovering interactions between modalities through time using a neural component called the Multi-attention Block (MAB) and storing them in the hybrid memory of a recurrent component called the Long-short Term Hybrid Memory (LSTHM). We perform extensive comparisons on six publicly available datasets for multimodal sentiment analysis, speaker trait recognition and emotion recognition. MARN shows state-of-the-art results performance in all the datasets.

Entities:  

Year:  2018        PMID: 32257595      PMCID: PMC7136010     

Source DB:  PubMed          Journal:  Proc Conf AAAI Artif Intell        ISSN: 2159-5399


  6 in total

1.  Functional and anatomical decomposition of face processing: evidence from prosopagnosia and PET study of normal subjects.

Authors:  J Sergent; J L Signoret
Journal:  Philos Trans R Soc Lond B Biol Sci       Date:  1992-01-29       Impact factor: 6.237

2.  Imaging first impressions: distinct neural processing of verbal and nonverbal social information.

Authors:  Bojana Kuzmanovic; Gary Bente; D Yves von Cramon; Leonhard Schilbach; Marc Tittgemeyer; Kai Vogeley
Journal:  Neuroimage       Date:  2011-12-27       Impact factor: 6.556

3.  Hidden conditional random fields.

Authors:  Ariadna Quattoni; Sybor Wang; Louis-Philippe Morency; Michael Collins; Trevor Darrell
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2007-10       Impact factor: 6.226

4.  Multiple Class Segmentation Using A Unified Framework over Mean-Shift Patches.

Authors:  Lin Yang; Peter Meer; David J Foran
Journal:  Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit       Date:  2007-07-16

5.  Long short-term memory.

Authors:  S Hochreiter; J Schmidhuber
Journal:  Neural Comput       Date:  1997-11-15       Impact factor: 2.026

6.  Neural synchronization during face-to-face communication.

Authors:  Jing Jiang; Bohan Dai; Danling Peng; Chaozhe Zhu; Li Liu; Chunming Lu
Journal:  J Neurosci       Date:  2012-11-07       Impact factor: 6.167

  6 in total
  10 in total

1.  Integrating Multimodal Information in Large Pretrained Transformers.

Authors:  Wasifur Rahman; Md Kamrul Hasan; Sangwu Lee; Amir Zadeh; Chengfeng Mao; Louis-Philippe Morency; Ehsan Hoque
Journal:  Proc Conf Assoc Comput Linguist Meet       Date:  2020-07

Review 2.  Multimodal deep learning for biomedical data fusion: a review.

Authors:  Sören Richard Stahlschmidt; Benjamin Ulfenborg; Jane Synnergren
Journal:  Brief Bioinform       Date:  2022-03-10       Impact factor: 11.622

3.  Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks.

Authors:  Zhibang Quan; Tao Sun; Mengli Su; Jishu Wei
Journal:  Comput Intell Neurosci       Date:  2022-08-09

4.  Forecasting the future clinical events of a patient through contrastive learning.

Authors:  Ziqi Zhang; Chao Yan; Xinmeng Zhang; Steve L Nyemba; Bradley A Malin
Journal:  J Am Med Inform Assoc       Date:  2022-08-16       Impact factor: 7.942

Review 5.  Harnessing multimodal data integration to advance precision oncology.

Authors:  Kevin M Boehm; Pegah Khosravi; Rami Vanguri; Jianjiong Gao; Sohrab P Shah
Journal:  Nat Rev Cancer       Date:  2021-10-18       Impact factor: 69.800

6.  Cross-Modal Sentiment Sensing with Visual-Augmented Representation and Diverse Decision Fusion.

Authors:  Sun Zhang; Bo Li; Chunyong Yin
Journal:  Sensors (Basel)       Date:  2021-12-23       Impact factor: 3.576

7.  LGCCT: A Light Gated and Crossed Complementation Transformer for Multimodal Speech Emotion Recognition.

Authors:  Feng Liu; Si-Yuan Shen; Zi-Wang Fu; Han-Yang Wang; Ai-Min Zhou; Jia-Yin Qi
Journal:  Entropy (Basel)       Date:  2022-07-21       Impact factor: 2.738

8.  AFR-BERT: Attention-based mechanism feature relevance fusion multimodal sentiment analysis model.

Authors:  Ji Mingyu; Zhou Jiawei; Wei Ning
Journal:  PLoS One       Date:  2022-09-09       Impact factor: 3.752

9.  Words Can Shift: Dynamically Adjusting Word Representations Using Nonverbal Behaviors.

Authors:  Yansen Wang; Ying Shen; Zhun Liu; Paul Pu Liang; Amir Zadeh; Louis-Philippe Morency
Journal:  Proc Conf AAAI Artif Intell       Date:  2019-07

10.  Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis.

Authors:  Jing He; Haonan Yanga; Changfan Zhang; Hongrun Chen; Yifu Xua
Journal:  Comput Intell Neurosci       Date:  2022-01-24
  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.