Automatic audiovisual integration in speech perception

被引:38
|
作者
Gentilucci, M [1 ]
Cattaneo, L [1 ]
机构
[1] Univ Parma, Dipartimento Neurosci, I-43100 Parma, Italy
关键词
McGurk effect; audiovisual integration; voice spectrum analysis; lip kinematics; imitation;
D O I
10.1007/s00221-005-0008-z
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Two experiments aimed to determine whether features of both the visual and acoustical inputs are always merged into the perceived representation of speech and whether this audiovisual integration is based on either cross-modal binding functions or on imitation. In a McGurk paradigm, observers were required to repeat aloud a string of phonemes uttered by an actor (acoustical presentation of phonemic string) whose mouth, in contrast, mimicked pronunciation of a different string (visual presentation). In a control experiment participants read the same printed strings of letters. This condition aimed to analyze the pattern of voice and the lip kinematics controlling for imitation. In the control experiment and in the congruent audiovisual presentation, i.e. when the articulation mouth gestures were congruent with the emission of the string of phones, the voice spectrum and the lip kinematics varied according to the pronounced strings of phonemes. In the McGurk paradigm the participants were unaware of the incongruence between visual and acoustical stimuli. The acoustical analysis of the participants' spoken responses showed three distinct patterns: the fusion of the two stimuli (the McGurk effect), repetition of the acoustically presented string of phonemes, and, less frequently, of the string of phonemes corresponding to the mouth gestures mimicked by the actor. However, the analysis of the latter two responses showed that the formant 2 of the participants' voice spectra always differed from the value recorded in the congruent audiovisual presentation. It approached the value of the formant 2 of the string of phonemes presented in the other modality, which was apparently ignored. The lip kinematics of the participants repeating the string of phonemes acoustically presented were influenced by the observation of the lip movements mimicked by the actor, but only when pronouncing a labial consonant. The data are discussed in favor of the hypothesis that features of both the visual and acoustical inputs always contribute to the representation of a string of phonemes and that cross-modal integration occurs by extracting mouth articulation features peculiar for the pronunciation of that string of phonemes.
引用
收藏
页码:66 / 75
页数:10
相关论文
共 50 条
  • [41] Audiovisual speech perception and its relation with temporal processing in children with and without autism
    Shuyuan Feng
    Haoyang Lu
    Jing Fang
    Xue Li
    Li Yi
    Lihan Chen
    Reading and Writing, 2023, 36 : 1419 - 1440
  • [42] Brief Report: Arrested Development of Audiovisual Speech Perception in Autism Spectrum Disorders
    Ryan A. Stevenson
    Justin K. Siemann
    Tiffany G. Woynaroski
    Brittany C. Schneider
    Haley E. Eberly
    Stephen M. Camarata
    Mark T. Wallace
    Journal of Autism and Developmental Disorders, 2014, 44 : 1470 - 1477
  • [43] Audiovisual Speech Perception in Children With Developmental Language Disorder in Degraded Listening Conditions
    Meronen, Auli
    Tiippana, Kaisa
    Westerholm, Jari
    Ahonen, Timo
    JOURNAL OF SPEECH LANGUAGE AND HEARING RESEARCH, 2013, 56 (01): : 211 - 221
  • [44] On the tip of the tongue: Modulation of the primary motor cortex during audiovisual speech perception
    Sato, Marc
    Buccino, Giovanni
    Gentilucci, Maurizio
    Cattaneo, Luigi
    SPEECH COMMUNICATION, 2010, 52 (06) : 533 - 541
  • [45] Brief Report: Arrested Development of Audiovisual Speech Perception in Autism Spectrum Disorders
    Stevenson, Ryan A.
    Siemann, Justin K.
    Woynaroski, Tiffany G.
    Schneider, Brittany C.
    Eberly, Haley E.
    Camarata, Stephen M.
    Wallace, Mark T.
    JOURNAL OF AUTISM AND DEVELOPMENTAL DISORDERS, 2014, 44 (06) : 1470 - 1477
  • [46] Electrophysiological evidence for a self-processing advantage during audiovisual speech integration
    Avril Treille
    Coriandre Vilain
    Sonia Kandel
    Marc Sato
    Experimental Brain Research, 2017, 235 : 2867 - 2876
  • [47] Electrophysiological evidence for a self-processing advantage during audiovisual speech integration
    Treille, Avril
    Vilain, Coriandre
    Kandel, Sonia
    Sato, Marc
    EXPERIMENTAL BRAIN RESEARCH, 2017, 235 (09) : 2867 - 2876
  • [48] Regularized models of audiovisual integration of speech with predictive power for sparse behavioral data
    Andersen, Tobias S.
    Winther, Ole
    JOURNAL OF MATHEMATICAL PSYCHOLOGY, 2020, 98
  • [49] Neurofunctional basis underlying audiovisual integration of print and speech sound in Chinese children
    Xia, Zhichao
    Yang, Ting
    Cui, Xin
    Hoeft, Fumiko
    Liu, Hong
    Zhang, Xianglin
    Shu, Hua
    Liu, Xiangping
    EUROPEAN JOURNAL OF NEUROSCIENCE, 2022, 55 (03) : 806 - 826
  • [50] Face-viewing patterns predict audiovisual speech integration in autistic children
    Feng, Shuyuan
    Lu, Haoyang
    Wang, Qiandong
    Li, Tianbi
    Fang, Jing
    Chen, Lihan
    Yi, Li
    AUTISM RESEARCH, 2021, 14 (12) : 2592 - 2602