Timing in audiovisual speech perception: A mini review and new psychophysical data

被引:17
|
作者
Venezia, Jonathan H. [1 ]
Thurman, Steven M. [2 ]
Matchin, William [3 ]
George, Sahara E. [4 ]
Hickok, Gregory [1 ]
机构
[1] Univ Calif Irvine, Dept Cognit Sci, Irvine, CA 92697 USA
[2] Univ Calif Los Angeles, Dept Psychol, Los Angeles, CA USA
[3] Univ Maryland, Dept Linguist, Baltimore, MD 21201 USA
[4] Univ Calif Irvine, Dept Anat & Neurobiol, Irvine, CA 92717 USA
关键词
Audiovisual speech; Multisensory integration; Prediction; Classification image; Timing; McGurk; Speech kinematics; MULTISENSORY INTEGRATION; VISUAL SPEECH; SPATIOTEMPORAL DYNAMICS; SUPERIOR COLLICULUS; MOVEMENT VELOCITY; TEMPORAL WINDOW; RECOGNITION; INFORMATION; TIME; SYNCHRONY;
D O I
10.3758/s13414-015-1026-y
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Recent influential models of audiovisual speech perception suggest that visual speech aids perception by generating predictions about the identity of upcoming speech sounds. These models place stock in the assumption that visual speech leads auditory speech in time. However, it is unclear whether and to what extent temporally-leading visual speech information contributes to perception. Previous studies exploring audiovisual-speech timing have relied upon psychophysical procedures that require artificial manipulation of cross-modal alignment or stimulus duration. We introduce a classification procedure that tracks perceptually relevant visual speech information in time without requiring such manipulations. Participants were shown videos of a McGurk syllable (auditory /apa/ + visual /aka/ = perceptual /ata/) and asked to perform phoneme identification (/apa/ yes-no). The mouth region of the visual stimulus was overlaid with a dynamic transparency mask that obscured visual speech in some frames but not others randomly across trials. Variability in participants' responses (similar to 35 % identification of /apa/ compared to similar to 5 % in the absence of the masker) served as the basis for classification analysis. The outcome was a high resolution spatiotemporal map of perceptually relevant visual features. We produced these maps for McGurk stimuli at different audiovisual temporal offsets (natural timing, 50-ms visual lead, and 100-ms visual lead). Briefly, temporally-leading (similar to 130 ms) visual information did influence auditory perception. Moreover, several visual features influenced perception of a single speech sound, with the relative influence of each feature depending on both its temporal relation to the auditory signal and its informational content.
引用
收藏
页码:583 / 601
页数:19
相关论文
共 50 条
  • [31] Audiovisual speech perception in Williams syndrome
    Böhning, M
    Campbell, R
    Karmiloff-Smith, A
    NEUROPSYCHOLOGIA, 2002, 40 (08) : 1396 - 1406
  • [32] Neural correlates of audiovisual speech synchrony perception and its relationship with autistic traits
    Zhou, Han-yu
    Zhang, Yi-jing
    Hu, Hui-xin
    Yan, Yong-jie
    Wang, Ling-ling
    Lui, Simon S. Y.
    Chan, Raymond C. K.
    PSYCH JOURNAL, 2023, 12 (04) : 514 - 523
  • [33] Brief Report: Arrested Development of Audiovisual Speech Perception in Autism Spectrum Disorders
    Stevenson, Ryan A.
    Siemann, Justin K.
    Woynaroski, Tiffany G.
    Schneider, Brittany C.
    Eberly, Haley E.
    Camarata, Stephen M.
    Wallace, Mark T.
    JOURNAL OF AUTISM AND DEVELOPMENTAL DISORDERS, 2014, 44 (06) : 1470 - 1477
  • [34] Audiovisual speech perception development at varying levels of perceptual processing
    Lalonde, Kaylah
    Holt, Rachael Frush
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2016, 139 (04) : 1713 - 1723
  • [35] Schizotypal traits are not related to multisensory integration or audiovisual speech perception
    Muller, Anne-Marie
    Dalal, Tyler C.
    Stevenson, Ryan A.
    CONSCIOUSNESS AND COGNITION, 2020, 86
  • [36] A possible neurophysiological correlate of audiovisual binding and unbinding in speech perception
    Ganesh, Attigodu C.
    Berthommier, Frederic
    Vilain, Coriandre
    Sato, Marc
    Schwartz, Jean-Luc
    FRONTIERS IN PSYCHOLOGY, 2014, 5
  • [37] Integration of audiovisual speech perception: From infancy to older adults
    Gijbels, Liesbeth
    Lee, Adrian K. C.
    Lalonde, Kaylah
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2025, 157 (03) : 1981 - 2000
  • [38] Increased Connectivity among Sensory and Motor Regions during Visual and Audiovisual Speech Perception
    Peelle, Jonathan E.
    Spehar, Brent
    Jones, Michael S.
    McConkey, Sarah
    Myerson, Joel
    Hale, Sandra
    Sommers, Mitchell S.
    Tye-Murray, Nancy
    JOURNAL OF NEUROSCIENCE, 2022, 42 (03) : 435 - 442
  • [39] A NETWORK ANALYSIS OF AUDIOVISUAL AFFECTIVE SPEECH PERCEPTION
    Jansma, H.
    Roebroeck, A.
    Munte, T. F.
    NEUROSCIENCE, 2014, 256 : 230 - 241
  • [40] Audiovisual perception of interrupted speech by nonnative listeners
    Yang, Jing
    Nagaraj, Naveen K.
    Magimairaj, Beula M.
    ATTENTION PERCEPTION & PSYCHOPHYSICS, 2024, 86 (05) : 1763 - 1776