Timing in audiovisual speech perception: A mini review and new psychophysical data

被引:17
|
作者
Venezia, Jonathan H. [1 ]
Thurman, Steven M. [2 ]
Matchin, William [3 ]
George, Sahara E. [4 ]
Hickok, Gregory [1 ]
机构
[1] Univ Calif Irvine, Dept Cognit Sci, Irvine, CA 92697 USA
[2] Univ Calif Los Angeles, Dept Psychol, Los Angeles, CA USA
[3] Univ Maryland, Dept Linguist, Baltimore, MD 21201 USA
[4] Univ Calif Irvine, Dept Anat & Neurobiol, Irvine, CA 92717 USA
关键词
Audiovisual speech; Multisensory integration; Prediction; Classification image; Timing; McGurk; Speech kinematics; MULTISENSORY INTEGRATION; VISUAL SPEECH; SPATIOTEMPORAL DYNAMICS; SUPERIOR COLLICULUS; MOVEMENT VELOCITY; TEMPORAL WINDOW; RECOGNITION; INFORMATION; TIME; SYNCHRONY;
D O I
10.3758/s13414-015-1026-y
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Recent influential models of audiovisual speech perception suggest that visual speech aids perception by generating predictions about the identity of upcoming speech sounds. These models place stock in the assumption that visual speech leads auditory speech in time. However, it is unclear whether and to what extent temporally-leading visual speech information contributes to perception. Previous studies exploring audiovisual-speech timing have relied upon psychophysical procedures that require artificial manipulation of cross-modal alignment or stimulus duration. We introduce a classification procedure that tracks perceptually relevant visual speech information in time without requiring such manipulations. Participants were shown videos of a McGurk syllable (auditory /apa/ + visual /aka/ = perceptual /ata/) and asked to perform phoneme identification (/apa/ yes-no). The mouth region of the visual stimulus was overlaid with a dynamic transparency mask that obscured visual speech in some frames but not others randomly across trials. Variability in participants' responses (similar to 35 % identification of /apa/ compared to similar to 5 % in the absence of the masker) served as the basis for classification analysis. The outcome was a high resolution spatiotemporal map of perceptually relevant visual features. We produced these maps for McGurk stimuli at different audiovisual temporal offsets (natural timing, 50-ms visual lead, and 100-ms visual lead). Briefly, temporally-leading (similar to 130 ms) visual information did influence auditory perception. Moreover, several visual features influenced perception of a single speech sound, with the relative influence of each feature depending on both its temporal relation to the auditory signal and its informational content.
引用
收藏
页码:583 / 601
页数:19
相关论文
共 50 条
  • [41] Visual and Audiovisual Effects of Isochronous Timing on Visual Perception and Brain Activity
    Marchant, Jennifer L.
    Driver, Jon
    CEREBRAL CORTEX, 2013, 23 (06) : 1290 - 1298
  • [42] Unimodal speech perception predicts stable individual differences in audiovisual benefit for phonemes, words and sentences
    von Seth, Jacqueline
    Aller, Mate
    Davis, Matthew H.
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2025, 157 (03) : 1554 - 1576
  • [43] Language/Culture Modulates Brain and Gaze Processes in Audiovisual Speech Perception
    Hisanaga, Satoko
    Sekiyama, Kaoru
    Igasaki, Tomohiko
    Murayama, Nobuki
    SCIENTIFIC REPORTS, 2016, 6
  • [44] Perception of Incongruent Audiovisual Speech: Distribution of Modality-Specific Responses
    Sandhya
    Vinay
    Manchaiah, V
    AMERICAN JOURNAL OF AUDIOLOGY, 2021, 30 (04) : 968 - 979
  • [45] A reanalysis of McGurk data suggests that audiovisual fusion in speech perception is subject-dependent
    Schwartz, Jean-Luc
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2010, 127 (03) : 1584 - 1594
  • [46] The contribution of dynamic visual cues to audiovisual speech perception
    Jaekl, Philip
    Pesquita, Ana
    Alsius, Agnes
    Munhall, Kevin
    Soto-Faraco, Salvador
    NEUROPSYCHOLOGIA, 2015, 75 : 402 - 410
  • [47] Reassessing the Benefits of Audiovisual Integration to Speech Perception and Intelligibility
    O'Hanlon, Brandon
    Plack, Christopher J.
    Nuttall, Helen E.
    JOURNAL OF SPEECH LANGUAGE AND HEARING RESEARCH, 2025, 68 (01): : 26 - 39
  • [48] Gaze behavior in audiovisual speech perception: The influence of ocular fixations on the McGurk effect
    Martin Paré
    Rebecca C. Richler
    Martin ten Hove
    K. G. Munhall
    Perception & Psychophysics, 2003, 65 : 553 - 567
  • [49] The effect of visual spatial attention on audiovisual speech perception in adults with Asperger syndrome
    Satu Saalasti
    Kaisa Tiippana
    Jari Kätsyri
    Mikko Sams
    Experimental Brain Research, 2011, 213 : 283 - 290
  • [50] The effect of visual spatial attention on audiovisual speech perception in adults with Asperger syndrome
    Saalasti, Satu
    Tiippana, Kaisa
    Katsyri, Jari
    Sams, Mikko
    EXPERIMENTAL BRAIN RESEARCH, 2011, 213 (2-3) : 283 - 290