Literature DB >> 33969362

CMU-MOSEAS: A Multimodal Language Dataset for Spanish, Portuguese, German and French.

Amir Zadeh1, Yan Sheng Cao2, Simon Hessner1, Paul Pu Liang3, Soujanya Poria4, Louis-Philippe Morency1.   

Abstract

Modeling multimodal language is a core research area in natural language processing. While languages such as English have relatively large multimodal language resources, other widely spoken languages across the globe have few or no large-scale datasets in this area. This disproportionately affects native speakers of languages other than English. As a step towards building more equitable and inclusive multimodal systems, we introduce the first large-scale multimodal language dataset for Spanish, Portuguese, German and French. The proposed dataset, called CMU-MOSEAS (CMU Multimodal Opinion Sentiment, Emotions and Attributes), is the largest of its kind with 40, 000 total labelled sentences. It covers a diverse set topics and speakers, and carries supervision of 20 labels including sentiment (and subjectivity), emotions, and attributes. Our evaluations on a state-of-the-art multimodal model demonstrates that CMU-MOSEAS enables further research for multilingual studies in multimodal language.

Entities:  

Year:  2020        PMID: 33969362      PMCID: PMC8106386          DOI: 10.18653/v1/2020.emnlp-main.141

Source DB:  PubMed          Journal:  Proc Conf Empir Methods Nat Lang Process


  6 in total

1.  Normalized amplitude quotient for parametrization of the glottal flow.

Authors:  Paavo Alku; Tom Bäckström; Erkki Vilkman
Journal:  J Acoust Soc Am       Date:  2002-08       Impact factor: 1.840

2.  Vocal intensity in speakers and singers.

Authors:  I R Titze; J Sundberg
Journal:  J Acoust Soc Am       Date:  1992-05       Impact factor: 1.840

3.  Vocal quality factors: analysis, synthesis, and perception.

Authors:  D G Childers; C K Lee
Journal:  J Acoust Soc Am       Date:  1991-11       Impact factor: 1.840

4.  Multimodal Transformer for Unaligned Multimodal Language Sequences.

Authors:  Yao-Hung Hubert Tsai; Shaojie Bai; Paul Pu Liang; J Zico Kolter; Louis-Philippe Morency; Ruslan Salakhutdinov
Journal:  Proc Conf Assoc Comput Linguist Meet       Date:  2019-07

5.  Mutual Correlation Attentive Factors in Dyadic Fusion Networks for Speech Emotion Recognition.

Authors:  Yue Gu; Xinyu Lyu; Weijia Sun; Weitian Li; Shuhong Chen; Xinyu Li; Marsic Ivan
Journal:  Proc ACM Int Conf Multimed       Date:  2019-10

6.  Why We Should Study Multimodal Language.

Authors:  Pamela Perniss
Journal:  Front Psychol       Date:  2018-06-28
  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.