Automatic audiovisual integration in speech perception

被引:38
|
作者
Gentilucci, M [1 ]
Cattaneo, L [1 ]
机构
[1] Univ Parma, Dipartimento Neurosci, I-43100 Parma, Italy
关键词
McGurk effect; audiovisual integration; voice spectrum analysis; lip kinematics; imitation;
D O I
10.1007/s00221-005-0008-z
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Two experiments aimed to determine whether features of both the visual and acoustical inputs are always merged into the perceived representation of speech and whether this audiovisual integration is based on either cross-modal binding functions or on imitation. In a McGurk paradigm, observers were required to repeat aloud a string of phonemes uttered by an actor (acoustical presentation of phonemic string) whose mouth, in contrast, mimicked pronunciation of a different string (visual presentation). In a control experiment participants read the same printed strings of letters. This condition aimed to analyze the pattern of voice and the lip kinematics controlling for imitation. In the control experiment and in the congruent audiovisual presentation, i.e. when the articulation mouth gestures were congruent with the emission of the string of phones, the voice spectrum and the lip kinematics varied according to the pronounced strings of phonemes. In the McGurk paradigm the participants were unaware of the incongruence between visual and acoustical stimuli. The acoustical analysis of the participants' spoken responses showed three distinct patterns: the fusion of the two stimuli (the McGurk effect), repetition of the acoustically presented string of phonemes, and, less frequently, of the string of phonemes corresponding to the mouth gestures mimicked by the actor. However, the analysis of the latter two responses showed that the formant 2 of the participants' voice spectra always differed from the value recorded in the congruent audiovisual presentation. It approached the value of the formant 2 of the string of phonemes presented in the other modality, which was apparently ignored. The lip kinematics of the participants repeating the string of phonemes acoustically presented were influenced by the observation of the lip movements mimicked by the actor, but only when pronouncing a labial consonant. The data are discussed in favor of the hypothesis that features of both the visual and acoustical inputs always contribute to the representation of a string of phonemes and that cross-modal integration occurs by extracting mouth articulation features peculiar for the pronunciation of that string of phonemes.
引用
收藏
页码:66 / 75
页数:10
相关论文
共 50 条
  • [31] High visual resolution matters in audiovisual speech perception, but only for some
    Alsius, Agnes
    Wayne, Rachel V.
    Pare, Martin
    Munhall, Kevin G.
    ATTENTION PERCEPTION & PSYCHOPHYSICS, 2016, 78 (05) : 1472 - 1487
  • [32] Perception of audiovisual speech synchrony for native and non-native language
    Navarra, Jordi
    Alsius, Agnes
    Velasco, Ignacio
    Soto-Faraco, Salvador
    Spence, Charles
    BRAIN RESEARCH, 2010, 1323 : 84 - 93
  • [33] The Effect of Extra-oral Facial Information on Audiovisual Speech Perception
    Huang, Zeyu
    Lu, Yao
    Wang, Lu
    Wu, Xiyu
    PROCEEDINGS OF THE 2018 INTERNATIONAL WORKSHOP ON EDUCATION REFORM AND SOCIAL SCIENCES (ERSS 2018), 2018, 300 : 808 - 813
  • [34] Auditory Event-Related Potentials (ERPs) in Audiovisual Speech Perception
    Pilling, Michael
    JOURNAL OF SPEECH LANGUAGE AND HEARING RESEARCH, 2009, 52 (04): : 1073 - 1081
  • [35] Effect of attentional load on audiovisual speech perception: evidence from ERPs
    Alsius, Agnes
    Moettoenen, Riikka
    Sams, Mikko E.
    Soto-Faraco, Salvador
    Tiippana, Kaisa
    FRONTIERS IN PSYCHOLOGY, 2014, 5
  • [36] The effect of face orientation on audiovisual speech integration in infancy: An electrophysiological study
    Szmytke, Magdalena
    Ilyka, Dianna
    Duda-Golawska, Joanna
    Laudanska, Zuzanna
    Malinowska-Korczak, Anna
    Tomalski, Przemyslaw
    DEVELOPMENTAL PSYCHOBIOLOGY, 2023, 65 (07)
  • [37] Increasing audiovisual speech integration in autism through enhanced attention to mouth
    Feng, Shuyuan
    Wang, Qiandong
    Hu, Yixiao
    Lu, Haoyang
    Li, Tianbi
    Song, Ci
    Fang, Jing
    Chen, Lihan
    Yi, Li
    DEVELOPMENTAL SCIENCE, 2023, 26 (04)
  • [38] Multisensory integration processes underlying speech perception as revealed by the McGurk illusion
    Marques, Lucas Murrins
    Lapenta, Olivia Morgan
    Costa, Thiago Leiros
    Boggio, Paulo Sergio
    LANGUAGE COGNITION AND NEUROSCIENCE, 2016, 31 (09) : 1115 - 1129
  • [39] Audiovisual speech perception and its relation with temporal processing in children with and without autism
    Feng, Shuyuan
    Lu, Haoyang
    Fang, Jing
    Li, Xue
    Yi, Li
    Chen, Lihan
    READING AND WRITING, 2023, 36 (06) : 1419 - 1440
  • [40] Gaze behavior in audiovisual speech perception: The influence of ocular fixations on the McGurk effect
    Martin Paré
    Rebecca C. Richler
    Martin ten Hove
    K. G. Munhall
    Perception & Psychophysics, 2003, 65 : 553 - 567