| Literature DB >> 15833302 |
Jyrki Tuomainen1, Tobias S Andersen, Kaisa Tiippana, Mikko Sams.
Abstract
In face-to-face conversation speech is perceived by ear and eye. We studied the prerequisites of audio-visual speech perception by using perceptually ambiguous sine wave replicas of natural speech as auditory stimuli. When the subjects were not aware that the auditory stimuli were speech, they showed only negligible integration of auditory and visual stimuli. When the same subjects learned to perceive the same auditory stimuli as speech, they integrated the auditory and visual stimuli in a similar manner as natural speech. These results demonstrate the existence of a multisensory speech-specific mode of perception.Mesh:
Year: 2004 PMID: 15833302 DOI: 10.1016/j.cognition.2004.10.004
Source DB: PubMed Journal: Cognition ISSN: 0010-0277