Automatic audiovisual integration in speech perception

被引:38
|
作者
Gentilucci, M [1 ]
Cattaneo, L [1 ]
机构
[1] Univ Parma, Dipartimento Neurosci, I-43100 Parma, Italy
关键词
McGurk effect; audiovisual integration; voice spectrum analysis; lip kinematics; imitation;
D O I
10.1007/s00221-005-0008-z
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Two experiments aimed to determine whether features of both the visual and acoustical inputs are always merged into the perceived representation of speech and whether this audiovisual integration is based on either cross-modal binding functions or on imitation. In a McGurk paradigm, observers were required to repeat aloud a string of phonemes uttered by an actor (acoustical presentation of phonemic string) whose mouth, in contrast, mimicked pronunciation of a different string (visual presentation). In a control experiment participants read the same printed strings of letters. This condition aimed to analyze the pattern of voice and the lip kinematics controlling for imitation. In the control experiment and in the congruent audiovisual presentation, i.e. when the articulation mouth gestures were congruent with the emission of the string of phones, the voice spectrum and the lip kinematics varied according to the pronounced strings of phonemes. In the McGurk paradigm the participants were unaware of the incongruence between visual and acoustical stimuli. The acoustical analysis of the participants' spoken responses showed three distinct patterns: the fusion of the two stimuli (the McGurk effect), repetition of the acoustically presented string of phonemes, and, less frequently, of the string of phonemes corresponding to the mouth gestures mimicked by the actor. However, the analysis of the latter two responses showed that the formant 2 of the participants' voice spectra always differed from the value recorded in the congruent audiovisual presentation. It approached the value of the formant 2 of the string of phonemes presented in the other modality, which was apparently ignored. The lip kinematics of the participants repeating the string of phonemes acoustically presented were influenced by the observation of the lip movements mimicked by the actor, but only when pronouncing a labial consonant. The data are discussed in favor of the hypothesis that features of both the visual and acoustical inputs always contribute to the representation of a string of phonemes and that cross-modal integration occurs by extracting mouth articulation features peculiar for the pronunciation of that string of phonemes.
引用
收藏
页码:66 / 75
页数:10
相关论文
共 50 条
  • [21] The Development of Cortical Responses to the Integration of Audiovisual Speech in Infancy
    Aleksandra A. W. Dopierała
    David López Pérez
    Evelyne Mercure
    Agnieszka Pluta
    Anna Malinowska-Korczak
    Samuel Evans
    Tomasz Wolak
    Przemysław Tomalski
    Brain Topography, 2023, 36 : 459 - 475
  • [22] Influence of language backgrounds on audiovisual speech perception across the lifespan
    Sekiyama, Kaoru
    ACOUSTICAL SCIENCE AND TECHNOLOGY, 2020, 41 (01) : 37 - 38
  • [23] AUDIOVISUAL SPEECH INTEGRATION IN THE SUPERIOR TEMPORAL REGION IS DYSFUNCTIONAL IN DYSLEXIA
    Ye, Zheng
    Ruesseler, Jascha
    Gerth, Ivonne
    Muente, Thomas F.
    NEUROSCIENCE, 2017, 356 : 1 - 10
  • [24] INVOLVEMENT OF SUPERIOR TEMPORAL AREAS IN AUDIOVISUAL AND AUDIOMOTOR SPEECH INTEGRATION
    Komeilipoor, N.
    Cesari, P.
    Daffertshofer, A.
    NEUROSCIENCE, 2017, 343 : 276 - 283
  • [25] Distinct cortical locations for integration of audiovisual speech and the McGurk effect
    Erickson, Laura C.
    Zielinski, Brandon A.
    Zielinski, Jennifer E. V.
    Liu, Guoying
    Turkeltaub, Peter E.
    Leaver, Amber M.
    Rauschecker, Josef P.
    FRONTIERS IN PSYCHOLOGY, 2014, 5
  • [26] Temporal synchrony and audiovisual integration of speech and object stimuli in autism
    Smith, Elizabeth
    Zhang, Shouling
    Bennetto, Loisa
    RESEARCH IN AUTISM SPECTRUM DISORDERS, 2017, 39 : 11 - 19
  • [27] Music expertise shapes audiovisual temporal integration windows for speech, sinewave speech, and music
    Lee, Hweeling
    Noppeney, Uta
    FRONTIERS IN PSYCHOLOGY, 2014, 5
  • [28] Increases in sensory noise predict attentional disruptions to audiovisual speech perception
    Fisher, Victoria L.
    Dean, Cassandra L.
    Nave, Claire S.
    Parkins, Emma V.
    Kerkhoff, Willa G.
    Kwakye, Leslie D.
    FRONTIERS IN HUMAN NEUROSCIENCE, 2023, 16
  • [29] Audiovisual Speech Perception and Eye Gaze Behavior of Adults with Asperger Syndrome
    Saalasti, Satu
    Katsyri, Jari
    Tiippana, Kaisa
    Laine-Hernandez, Mari
    von Wendt, Lennart
    Sams, Mikko
    JOURNAL OF AUTISM AND DEVELOPMENTAL DISORDERS, 2012, 42 (08) : 1606 - 1615
  • [30] High visual resolution matters in audiovisual speech perception, but only for some
    Agnès Alsius
    Rachel V. Wayne
    Martin Paré
    Kevin G. Munhall
    Attention, Perception, & Psychophysics, 2016, 78 : 1472 - 1487