Literature DB >> 33824190

Neurophysiological indices of audiovisual speech processing reveal a hierarchy of multisensory integration effects.

Aisling E O'Sullivan1, Michael J Crosse2, Giovanni M Di Liberto3, Alain de Cheveigné3,4, Edmund C Lalor5,6.   

Abstract

Seeing a speaker's face benefits speech comprehension, especially in challenging listening conditions. This perceptual benefit is thought to stem from the neural integration of visual and auditory speech at multiple stages of processing, whereby movement of a speaker's face provides temporal cues to auditory cortex, and articulatory information from the speaker's mouth can aid recognizing specific linguistic units (e.g., phonemes, syllables). However it remains unclear how the integration of these cues varies as a function of listening conditions. Here we sought to provide insight on these questions by examining EEG responses in humans (males and females) to natural audiovisual, audio, and visual speech in quiet and in noise. We represented our speech stimuli in terms of their spectrograms and their phonetic features, and then quantified the strength of the encoding of those features in the EEG using canonical correlation analysis. The encoding of both spectrotemporal and phonetic features was shown to be more robust in audiovisual speech responses then what would have been expected from the summation of the audio and visual speech responses, suggesting that multisensory integration occurs at both spectrotemporal and phonetic stages of speech processing. We also found evidence to suggest that the integration effects may change with listening conditions, however this was an exploratory analysis and future work will be required to examine this effect using a within-subject design. These findings demonstrate that integration of audio and visual speech occurs at multiple stages along the speech processing hierarchy.SIGNIFICANCE STATEMENT During conversation, visual cues impact our perception of speech. Integration of auditory and visual speech is thought to occur at multiple stages of speech processing and vary flexibly depending on the listening conditions. Here we examine audiovisual integration at two stages of speech processing using the speech spectrogram and a phonetic representation, and test how audiovisual integration adapts to degraded listening conditions. We find significant integration at both of these stages regardless of listening conditions. These findings reveal neural indices of multisensory interactions at different stages of processing and provide support for the multistage integration framework.
Copyright © 2021 the authors.

Entities:  

Year:  2021        PMID: 33824190     DOI: 10.1523/JNEUROSCI.0906-20.2021

Source DB:  PubMed          Journal:  J Neurosci        ISSN: 0270-6474            Impact factor:   6.167


  7 in total

1.  Binding the Acoustic Features of an Auditory Source through Temporal Coherence.

Authors:  Mohsen Rezaeizadeh; Shihab Shamma
Journal:  Cereb Cortex Commun       Date:  2021-10-06

Review 2.  Linear Modeling of Neurophysiological Responses to Speech and Other Continuous Stimuli: Methodological Considerations for Applied Research.

Authors:  Michael J Crosse; Nathaniel J Zuk; Giovanni M Di Liberto; Aaron R Nidiffer; Sophie Molholm; Edmund C Lalor
Journal:  Front Neurosci       Date:  2021-11-22       Impact factor: 4.677

3.  Editorial: Neural Tracking: Closing the Gap Between Neurophysiology and Translational Medicine.

Authors:  Giovanni M Di Liberto; Jens Hjortkjær; Nima Mesgarani
Journal:  Front Neurosci       Date:  2022-03-16       Impact factor: 5.152

4.  Intelligibility of audiovisual sentences drives multivoxel response patterns in human superior temporal cortex.

Authors:  Johannes Rennig; Michael S Beauchamp
Journal:  Neuroimage       Date:  2021-12-11       Impact factor: 6.556

5.  Enhancement of speech-in-noise comprehension through vibrotactile stimulation at the syllabic rate.

Authors:  Pierre Guilleminot; Tobias Reichenbach
Journal:  Proc Natl Acad Sci U S A       Date:  2022-03-21       Impact factor: 12.779

6.  Neurosensory development of the four brainstem-projecting sensory systems and their integration in the telencephalon.

Authors:  Bernd Fritzsch; Karen L Elliott; Ebenezer N Yamoah
Journal:  Front Neural Circuits       Date:  2022-09-23       Impact factor: 3.342

7.  Speech-Driven Facial Animations Improve Speech-in-Noise Comprehension of Humans.

Authors:  Enrico Varano; Konstantinos Vougioukas; Pingchuan Ma; Stavros Petridis; Maja Pantic; Tobias Reichenbach
Journal:  Front Neurosci       Date:  2022-01-05       Impact factor: 4.677

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.