Literature DB >> 35845168

Audition controls the flow of visual time during multisensory perception.

Mariel G Gonzales1, Kristina C Backer1,2, Yueqi Yan2, Lee M Miller3,4,5, Heather Bortfeld1,2,6, Antoine J Shahin1,2.   

Abstract

Previous work addressing the influence of audition on visual perception has mainly been assessed using non-speech stimuli. Herein, we introduce the Audiovisual Time-Flow Illusion in spoken language, underscoring the role of audition in multisensory processing. When brief pauses were inserted into or brief portions were removed from an acoustic speech stream, individuals perceived the corresponding visual speech as "pausing" or "skipping", respectively-even though the visual stimulus was intact. When the stimulus manipulation was reversed-brief pauses were inserted into, or brief portions were removed from the visual speech stream-individuals failed to perceive the illusion in the corresponding intact auditory stream. Our findings demonstrate that in the context of spoken language, people continually realign the pace of their visual perception based on that of the auditory input. In short, the auditory modality sets the pace of the visual modality during audiovisual speech processing.
© 2022 The Author(s).

Entities:  

Keywords:  Biological sciences; neuroscience; sensory neuroscience

Year:  2022        PMID: 35845168      PMCID: PMC9283509          DOI: 10.1016/j.isci.2022.104671

Source DB:  PubMed          Journal:  iScience        ISSN: 2589-0042


Introduction

The visual modality’s influence on auditory perception of spoken language is well established. A classic example of this is the McGurk illusion, wherein the incongruent pairing of visual and auditory speech results in the auditory perception of either the visually conveyed phoneme or a different, third phoneme (McGurk and MacDonald, 1976; Abbott and Shahin, 2018; Shahin et al., 2018). However, the impact of the auditory modality on visual perception in spoken language processing is less understood. Examples from the non-speech domain (Welch and Warren, 1980; Recanzone, 2003; Stein et al., 1996; Sekuler et al., 1997; Shams et al., 2000, 2001, 2002; Vroomen and de Gelder, 2000, 2004; Chen and Vroomen, 2013) suggest a critical role for audition in shaping visual perception. For example, in the Double Flash Illusion (Shams et al., 2000, 2001, 2002) when one flash is paired with two tones, individuals often see two flashes. However, when two flashes are paired with one tone, individuals often see one flash. In the flash-lag effect (FLE), perception of a flash of light lags relative to a moving object depending on when bursts of sound are played. The FLE was found to vary linearly with when the bursts of sound occur, that is, either before, during, or after each flash (Vroomen and de Gelder, 2004). These findings suggest that in audiovisual crossmodal perception, the auditory modality is used as a temporal reference for the visual modality. However, the above-mentioned illusions cannot address two crucial questions: i) whether audition exerts such an influence over vision in the context of complex, real world multisensory stimuli such as audiovisual speech, and ii) whether audition merely serves as a temporal reference for vision, or whether it actually controls the flow of perceived visual time to maintain crossmodal alignment. We posited that the role of the auditory modality is to set the pace of the visual modality by controlling (interrupting or modulating) perceived visual time, so that visual temporal processing synchronizes with the information conveyed by audition. In the experimental blocks, individuals attended to videos of a speaker uttering trisyllabic words. In one condition, the acoustic speech signal had two silent segments inserted (A-pause), inducing a perception that the acoustic speech was “pausing” or slowing-down; in another condition, the audio had two speech segments excised (A-skip), inducing a perception that the acoustic speech was “skipping” or speeding-up. Figure 1 depicts example waveforms and spectrograms of manipulated audios. (Note that we use quotes to indicate subjective perception, whereas an absence of quotes denotes physical stimulus properties.) The visual part of the video remained unchanged. The experimental blocks also included intact audiovisual stimuli (unchanged) and catch trials whereby both the audio and visual parts of the videos were pausing (AV-pause) or skipping (AV-skip) (i.e., temporally aligned segments were inserted or excised across both modalities). Individuals made judgments about whether the visual part of the video was “pausing” (“V-pause”), “skipping” (“V-skip”), or remained “unchanged” (“V-unchanged”). In the control blocks, the stimulus manipulations were reversed. That is, individuals listened to intact audio streams paired with manipulated visual streams (V-skip, V-pause or unchanged) while making judgments about whether the auditory part was “pausing” (“A-pause”), “skipping” (“A-skip”), or remained “unchanged” (“A-unchanged”). As in the experimental blocks, the control blocks also included intact and catch trials. Within each block, the stimuli were randomly intermixed, and the order of the experimental and control blocks was counterbalanced across subjects.
Figure 1

Example audio waveforms and spectrograms

(A) Example audio waveforms of two manipulated words, that led to induction of “skipping” and “pausing” perception in the visual modality.

Example audio spectrograms of the two manipulated words shown in (A).

Example audio waveforms and spectrograms (A) Example audio waveforms of two manipulated words, that led to induction of “skipping” and “pausing” perception in the visual modality. Example audio spectrograms of the two manipulated words shown in (A). We hypothesized that individuals would perceive more robust “pausing” or “skipping” in the visual speech (“V-pause” or “V-skip”) when the auditory speech paused or skipped (A-pause or A-skip), respectively, relative to the opposite manipulation—judging intact acoustic stimuli as “A-pause” or “A-skip,” when the visual stimuli paused or skipped (“V-pause” or “V-skip”), respectively. Such a finding would indicate that the auditory modality sets the pace of the visual modality, and not the other way around.

Results

Figure 2 (see also Table S1: Mean response percentages and standard deviations, relating to Figure 2) depicts group percentages for “pause” and “skip” illusory percepts, as well as “unchanged” percepts. These percentages were calculated as the number of trials perceived as either “pause,” “skip,” or “unchanged” in one modality, divided by all trials within the modality (see examples in Figure 2 and Table S1). In the experimental blocks, the average percentage of individuals who experienced a “pause” illusion in the intact visual stream (“V-pause”) because of a physical pause in the auditory stream (A-pause) was 35.4%, significantly different from the percentage experiencing a “pause” illusion in the unchanged trials (5.3%, p < 0.001). Participants perceived a “skip” illusion in the intact visual stream (“V-skip”) because of a physical skip in the auditory stream (A-skip) 48% of the time, which was also significantly different from the average probability of experiencing a “skip” illusion in the unchanged trials (2.2%, p < 0.001). In the control blocks, individuals experienced an “A-pause” or “A-skip” illusion in the intact auditory stream because of a pause or skip in the visual stream (V-pause or V-skip) 18.1 and 14.7% of the time, respectively. Both percentages were significantly different from the same percentages for the unchanged trials (V-pause” perceived as “A-pause” vs. unchanged perceived as “A-pause”: 18.1 vs. 9.6%, p < 0.001; V-skip perceived as “A-skip” vs. unchanged perceived as “A-skip”: 14.7 vs. 1.5%, p < 0.001).
Figure 2

Box plots of response (Percept) percentages in the auditory (A) or visual (B)modality induced by stimulus manipulation in the counter modality

For example, the percentage of the percept “A-pause” in response to a stimulus of V-pause indicates the percentage that individuals perceived a “pause” in the auditory stream in responses to a physical pause in the visual stream that is paired with an acoustically intact auditory stream.

Box plots of response (Percept) percentages in the auditory (A) or visual (B)modality induced by stimulus manipulation in the counter modality For example, the percentage of the percept “A-pause” in response to a stimulus of V-pause indicates the percentage that individuals perceived a “pause” in the auditory stream in responses to a physical pause in the visual stream that is paired with an acoustically intact auditory stream. We next conducted mixed effects logistic regression analysis to test our hypothesis directly—whether the illusions observed in the visual stream because of auditory stimulus alterations were more robust than the illusions observed in the auditory stream because of visual stimulus alterations. Results demonstrated a significant difference in the likelihood of individuals experiencing the “pause” or “skip” illusions in the visual stream when the alterations occurred in the auditory streams, versus experiencing the illusions in the auditory stream when the alterations occurred in the visual streams [for A-pause perceived as “V-pause” vs. V-pause perceived as “A-pause”, [“V-Pause” vs. "A-Pause", Odds ratio (OR) = 0.36 [0.29, 0.44], p < 0.001; for “V-Skip” vs. "A-Skip", OR = 0.17 [0.14, 0.20], p < 0.001]. Finally, we evaluated the differences in unchanged trials (no change in either audiovisual stream) occurring in the experimental and controls conditions and catch (AV-pause, AV-skip) trials (temporal manipulation in both modalities) occurring in the experimental and control conditions, to ascertain whether participants across both tasks were performing comparably. No significant change was observed across the main and control blocks for the catch trials (for AV-pause perceived as “V-pause” in main versus AV-pause perceived as “A-pause” in control, OR: 1.23[0.70, 2.17], p = 0.471; for AV-skip perceived as “V-skip” in main vs. AV-skip perceived as “A-skip” in control, OR: 1.65[0.61, 4.47], p = 0.324). However, in the unchanged conditions, performance differed significantly between the experimental and control blocks. Despite the means being close to the same (“V-unchanged” = 92.5%; “A-unchanged” = 89%), participants were more accurate in labeling the intact unchanged streams as “V-unchanged” (main blocks) than the intact unchanged streams as “A-unchanged” (control blocks) [for unchanged correctly perceived in main vs. control, OR: 0.64[0.48, 0.85], p < 0.001]. In short, participants’ comparable performance on the unchanged (despite a significant difference) and catch trials across the experimental and control blocks demonstrates that participants were vigilant in their performance of the tasks. These results reinforce the findings observed during the unimodal “skip” and “pause” manipulations, lending additional support to the interpretation that the illusory effects observed between the experimental relative to the control blocks are true perceptual phenomena.

Discussion

Our results show that the auditory modality, not the visual modality, sets the pace of audiovisual speech processing, whereby the visual modality “skips” and “pauses” to keep pace with the rate of information conveyed by the auditory modality. This is in line with a temporal reference role for the auditory modality in multisensory processing, but goes well beyond it, showing that audition actually controls the flow of time for visually perceived events. In ecologically appropriate situations, sounds often lead us to redirect our visual attention, with the visual modality recalibrating its spatial focus. In other circumstances, e.g., discourse, rather than recalibrating its spatial focus, vision recalibrates its temporal focus as demonstrated here. The current findings raise the question of why the visual modality synchronizes with the pace of the auditory input, but not the reverse. In natural discourse, the visual modality (mouth movement) often precedes corresponding sound production (Chandrasekaran et al., 2009; Schwartz and Savariaux, 2014) and provides predictive cues to optimize the processing of sound. Thus, visual input plays a major role, behaviorally and neurophysiologically, in guiding our predictions about the unfolding speech signal (Besle et al., 2004; van Wassenhove et al., 2005). This predictive quality enhances speech comprehension, especially in noisy situations (Sumby and Pollack, 1954). Here we provide strong complementary evidence that audition, the more temporally precise modality, controls the flow of time in vision, the more temporally predictive modality. This raises important mechanistic and phenomenological questions about neural computational time versus experienced time, not only across sensory modalities but in sensorimotor time as well (De Kock et al., 2021). In view of the current findings, we posit that as the speech stream changes pace, the visual modality realigns its pace according to the auditory modality to reassume a leading position. In this way, the visual modality maintains its predictive impact on audition and optimizes speech intelligibility.

Limitations of the study

Our understanding of the Audiovisual Time-Flow illusion remains limited with regard to the neural and contextual mechanisms that drive it. For example, it is unclear whether the effect is associated with transient changes in the auditory stream as in the present experiment, or if it is also observed during a slow buildup of audiovisual stream misalignments, such as when the audio is sped up or slowed down. A second limitation is whether this phenomenon is speech specific, a question that should be addressed through the use of nonspeech stimuli, such as a musical performance or a movie clip of an inanimate object’s movement. A third limitation, building on the previous two, is whether this phenomenon is driven by low-level correspondence between the auditory and visual streams or is linguistically driven. For example, if individuals can still experience the illusion with incongruent speech streams, whereby phonemes and visemes are misaligned, one can conclude that the illusion is driven by low level correspondence of AV streams as opposed to higher level linguistic factors. Finally, it is imperative to assess the neurophysiology that facilitates the transfer of synchrony from the auditory modality to the visual modality to determine whether this transfer is directly communicated between nominally unisensory cortices or mediated by a higher-order multisensory hub.

STAR★Methods

Key resources table

Resource availability

Lead contact

Questions and requests for information and data/code should be directed to the corresponding author (ashahin@ucmerced.edu).

Materials availability

Stimulus examples and original raw behavioral response data as text files can be found on https://data.mendeley.com/datasets/sxrchpgvpg/1. The site includes a text file explaining the variables in the text files.

Experimental model and subject details

Participants

Twenty-four individuals (>18 years of age, M = 21.25 years, SD = 4.33; 16 females) participated in the study. All participants reported native/fluent English background, normal hearing, right-handedness, and normal/corrected-to-normal vision. Only one participant was a non-native English speaker, who reported that they began learning English before the age 10 years. Prior to participation, participants provided written informed consent, and filled out a questionnaire about their language background, educational level, handedness and neurological history. Informed consent was obtained from the participants prior to participation, and Experimental methods were carried out in accordance with protocols approved by the Institutional Review Board (IRB) of the University of California, Merced. Monetary compensation was given to all participants for their participation.

Method details

Stimuli experimental blocks

The stimuli consisted of 3-s audiovisual video clips created by merging silent videos and corresponding audio clips of a female talker uttering trisyllabic words. The visual portions of the videos were cropped to only show the talker’s face below the eyes to the bottom of the neck. There were three audiovisual stimulus conditions: unchanged, paused, skipped. Each condition consisted of 50 videos. To control for context influences, the words were different across the three conditions. The visual portions of the videos of all three conditions were intact. The three conditions only varied in the acoustic portions of the videos. In the unchanged condition, the audio portions of the words were intact, uttered in normal form. In the paused condition, the audio portions of the words contained two 100-ms silent segments inserted in the audio 200-ms apart, giving the perception that the audio is pausing. The onset of the first silence insertion began about a third of the way into the beginning of the word, followed by a 200-ms of the remaining speech signal, followed by another 100-ms silence, and finally followed by the rest of the speech signal. These parameters were based on the senior author’s own judgment, as to what parameters might induce the strongest pause illusion. In the skipped condition, two speech segments were excised from the audios to induce a perception that the audio is skipping. A 100-ms segment was excised, beginning third of the way of the word, followed by a 133-ms of the speech, followed by another 100-ms excision, and finally followed by the rest of the word. Again, these arbitrary parameters were based on the senior author’s judgment. Finally, we also included 5 catch trials in which the audio and corresponding visual portions of the videos were excised (AV-skip), and 5 catch trials in which the audio contained silent portions and the visual portions contained paused (repeated frames) corresponding to the same time windows of the audio’s silent segments (AV-pause). The words used in the catch trials were different from the original three conditions. The reason for inclusion of the catch trials, was to monitor the participants’ task vigilance. All audio clips were group-normalized in Adobe Audition to the same sound pressure level of −25 dB and presented at an average of 56 dB SPL (measured at head location). The audios can be accessed via the following link: https://data.mendeley.com/datasets/sxrchpgvpg/1.

Control blocks

The manipulations of the audiovisual stimuli in the control blocks were reversed across the two modalities for the three original conditions. The same words were used, as well as catch trial conditions.

Procedure

The participants sat about a meter from a 27-inch monitor and performed the task in a sound-attenuated booth. The task consisted of 5 blocks: Two practice blocks, 2 experimental blocks, and 2 control blocks. The practice blocks, one for the main and one for the control, consisted of 6 trials each: 2 unchanged, 2 paused, and 2 skipped trials. The main and control blocks each consisted of 75 randomly ordered trials for a total of 150 trials: 50 unchanged, 50 paused, and 50 skipped trials. The main and control block order was counterbalanced across subjects—12 participants received the main blocks first, and the other 12 participants received the control blocks first. Visual stimuli were presented through the 27-inch monitor, and auditory stimuli were presented through two loudspeakers located on the left and right sides (−/+45°) of the monitor. The sound was heard as coming directly from midline (0°). The participants were instructed to indicate whether they see the visual portion of the videos as skipping or speeding-up, pausing (“freezing”) or “slowing-down”, or neither skipping nor pausing (unchanged), by pressing the left arrow, right arrow, or down arrow, respectively, on the keyboard.

Data analysis

Logfiles of the participants’ responses were parsed in MATLAB 2015 (MathWorks, Natick MA). Individual mean percept percentages were calculated as the number of responses for each percept (e.g., “skip”) divided by all responses within each stimulus type (e.g., skip) x 100. Since there were three response options, chance level was therefore 33%.

Quantification and statistical analysis

We originally planned to run analysis of variance (ANOVA) to examine the percentage difference across modalities and conditions and run pairwise comparisons, but due to the violation of analysis of variance (ANOVA) assumptions of independence of the data, e.g., the percentage of A-pause as “V-pause” plus A-pause as “V-skip” plus A-pause as “V-unchanged” is equal to 1 (see Penn State University. Applied Statistics. 10.2.1 - ANOVA Assumptions. Retrieved 5/25/2022 From: https://online.stat.psu.edu/stat500/lesson/10/10.2/10.2.1#:∼:text=There%20are%20three%20primary%20assumptions,The%20data%20are%20independent), we decided against an ANOVA. Rather, Separate bootstrap t-tests were performed to test whether the average percentage for each illusory percept of interests was significantly greater or lower than the average percentage for the same illusory percept in the unchanged condition, accounting for non-normality of the population of measurements (Kohl, 2020). Observations were considered dependent within individuals. Therefore, mixed-effects logistic regression models were computed on the single-trial data (i.e., observations) and used to compare the intact conditions or catch trials across the main and control blocks (i.e., fixed effects), while allowing a random intercept to account for the dependence of observations within individuals. The R package “MKinfer “(Kohl, 2020) and “mclogit” (Elff, 2021) were used to perform the bootstrap t-tests and mixed-effects models, respectively.
REAGENT or RESOURCESOURCEIDENTIFIER
Deposited data

Raw Behavioral Responseshttps://data.mendeley.com/datasets/sxrchpgvpg/1.
Example Stimulihttps://data.mendeley.com/datasets/sxrchpgvpg/1.

Software and algorithms

In-house stimulus presentation codehttps://www.neurobs.com/
In-house Matlab parse codehttps://www.mathworks.com/products/matlab.html
  18 in total

1.  Illusions. What you see is what you hear.

Authors:  L Shams; Y Kamitani; S Shimojo
Journal:  Nature       Date:  2000-12-14       Impact factor: 49.962

2.  Sound enhances visual perception: cross-modal effects of auditory organization on vision.

Authors:  J Vroomen; B de Gelder
Journal:  J Exp Psychol Hum Percept Perform       Date:  2000-10       Impact factor: 3.332

3.  Visual speech speeds up the neural processing of auditory speech.

Authors:  Virginie van Wassenhove; Ken W Grant; David Poeppel
Journal:  Proc Natl Acad Sci U S A       Date:  2005-01-12       Impact factor: 11.205

Review 4.  Intersensory binding across space and time: a tutorial review.

Authors:  Lihan Chen; Jean Vroomen
Journal:  Atten Percept Psychophys       Date:  2013-07       Impact factor: 2.199

5.  Enhancement of perceived visual intensity by auditory stimuli: a psychophysical analysis.

Authors:  B E Stein; N London; L K Wilkinson; D D Price
Journal:  J Cogn Neurosci       Date:  1996-11       Impact factor: 3.225

6.  Sound alters visual motion perception.

Authors:  R Sekuler; A B Sekuler; R Lau
Journal:  Nature       Date:  1997-01-23       Impact factor: 49.962

7.  Neural Mechanisms Underlying Cross-Modal Phonetic Encoding.

Authors:  Antoine J Shahin; Kristina C Backer; Lawrence D Rosenblum; Jess R Kerlin
Journal:  J Neurosci       Date:  2017-12-20       Impact factor: 6.167

Review 8.  Immediate perceptual response to intersensory discrepancy.

Authors:  R B Welch; D H Warren
Journal:  Psychol Bull       Date:  1980-11       Impact factor: 17.737

Review 9.  How movements shape the perception of time.

Authors:  Rose De Kock; Keri Anne Gladhill; Minaz Numa Ali; Wilsaan Mychal Joiner; Martin Wiener
Journal:  Trends Cogn Sci       Date:  2021-09-13       Impact factor: 20.229

10.  Auditory influences on visual temporal rate perception.

Authors:  Gregg H Recanzone
Journal:  J Neurophysiol       Date:  2003-02       Impact factor: 2.714

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.