Literature DB >> 26669309

Timing in audiovisual speech perception: A mini review and new psychophysical data.

Jonathan H Venezia1, Steven M Thurman2, William Matchin3, Sahara E George4, Gregory Hickok5.   

Abstract

Recent influential models of audiovisual speech perception suggest that visual speech aids perception by generating predictions about the identity of upcoming speech sounds. These models place stock in the assumption that visual speech leads auditory speech in time. However, it is unclear whether and to what extent temporally-leading visual speech information contributes to perception. Previous studies exploring audiovisual-speech timing have relied upon psychophysical procedures that require artificial manipulation of cross-modal alignment or stimulus duration. We introduce a classification procedure that tracks perceptually relevant visual speech information in time without requiring such manipulations. Participants were shown videos of a McGurk syllable (auditory /apa/ + visual /aka/ = perceptual /ata/) and asked to perform phoneme identification (/apa/ yes-no). The mouth region of the visual stimulus was overlaid with a dynamic transparency mask that obscured visual speech in some frames but not others randomly across trials. Variability in participants' responses (~35 % identification of /apa/ compared to ~5 % in the absence of the masker) served as the basis for classification analysis. The outcome was a high resolution spatiotemporal map of perceptually relevant visual features. We produced these maps for McGurk stimuli at different audiovisual temporal offsets (natural timing, 50-ms visual lead, and 100-ms visual lead). Briefly, temporally-leading (~130 ms) visual information did influence auditory perception. Moreover, several visual features influenced perception of a single speech sound, with the relative influence of each feature depending on both its temporal relation to the auditory signal and its informational content.

Entities:  

Keywords:  Audiovisual speech; Classification image; McGurk; Multisensory integration; Prediction; Speech kinematics; Timing

Mesh:

Year:  2016        PMID: 26669309      PMCID: PMC4744562          DOI: 10.3758/s13414-015-1026-y

Source DB:  PubMed          Journal:  Atten Percept Psychophys        ISSN: 1943-3921            Impact factor:   2.199


  75 in total

1.  Evidence from functional magnetic resonance imaging of crossmodal binding in the human heteromodal cortex.

Authors:  G A Calvert; R Campbell; M J Brammer
Journal:  Curr Biol       Date:  2000-06-01       Impact factor: 10.834

2.  Neural correlates of multisensory integration of ecologically valid audiovisual events.

Authors:  Jeroen J Stekelenburg; Jean Vroomen
Journal:  J Cogn Neurosci       Date:  2007-12       Impact factor: 3.225

Review 3.  The processing of audio-visual speech: empirical and neural bases.

Authors:  Ruth Campbell
Journal:  Philos Trans R Soc Lond B Biol Sci       Date:  2008-03-12       Impact factor: 6.237

4.  Audiovisual gating and the time course of speech perception.

Authors:  K G Munhall; Y Tohkura
Journal:  J Acoust Soc Am       Date:  1998-07       Impact factor: 1.840

5.  Auditory-visual speech perception and synchrony detection for speech and nonspeech signals.

Authors:  Brianna Conrey; David B Pisoni
Journal:  J Acoust Soc Am       Date:  2006-06       Impact factor: 1.840

6.  Hearing by eye.

Authors:  R Campbell; B Dodd
Journal:  Q J Exp Psychol       Date:  1980-02       Impact factor: 2.143

7.  The natural statistics of audiovisual speech.

Authors:  Chandramouli Chandrasekaran; Andrea Trubanova; Sébastien Stillittano; Alice Caplier; Asif A Ghazanfar
Journal:  PLoS Comput Biol       Date:  2009-07-17       Impact factor: 4.475

8.  Similarity structure in visual speech perception and optical phonetic signals.

Authors:  Jintao Jiang; Edward T Auer; Abeer Alwan; Patricia A Keating; Lynne E Bernstein
Journal:  Percept Psychophys       Date:  2007-10

9.  Diagnostic spatial frequencies and human efficiency for discriminating actions.

Authors:  Steven M Thurman; Emily D Grossman
Journal:  Atten Percept Psychophys       Date:  2011-02       Impact factor: 2.199

10.  Temporal structure and complexity affect audio-visual correspondence detection.

Authors:  Rachel N Denison; Jon Driver; Christian C Ruff
Journal:  Front Psychol       Date:  2013-01-22
View more
  3 in total

1.  Surmising synchrony of sound and sight: Factors explaining variance of audiovisual integration in hurdling, tap dancing and drumming.

Authors:  Nina Heins; Jennifer Pomp; Daniel S Kluger; Stefan Vinbrüx; Ima Trempler; Axel Kohler; Katja Kornysheva; Karen Zentgraf; Markus Raab; Ricarda I Schubotz
Journal:  PLoS One       Date:  2021-07-22       Impact factor: 3.240

2.  Comparison of McGurk Effect across Three Consonant-Vowel Combinations in Kannada.

Authors:  Dhatri S Devaraju; Ajith Kumar U; Santosh Maruthy
Journal:  J Audiol Otol       Date:  2018-12-07

3.  Auditory and visual distractors disrupt multisensory temporal acuity in the crossmodal temporal order judgment task.

Authors:  Cassandra L Dean; Brady A Eggleston; Kyla David Gibney; Enimielen Aligbe; Marissa Blackwell; Leslie Dowell Kwakye
Journal:  PLoS One       Date:  2017-07-19       Impact factor: 3.240

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.