Audio-visual speech scene analysis: Characterization of the dynamics of unbinding and rebinding the McGurk effect

被引:25
|
作者
Nahorna, Olha [1 ]
Berthommier, Frederic [1 ]
Schwartz, Jean-Luc [1 ]
机构
[1] Grenoble Univ, CNRS, Speech & Cognit Dept, GIPSA Lab,UMR 5216, Grenoble, France
基金
欧洲研究理事会;
关键词
VISUAL SPEECH; SPATIAL ATTENTION; AUDITORY SPEECH; BIMODAL SPEECH; PERCEPTION; INTEGRATION; INFORMATION; DECISIONS; VOICES; INTELLIGIBILITY;
D O I
10.1121/1.4904536
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
While audiovisual interactions in speech perception have long been considered as automatic, recent data suggest that this is not the case. In a previous study, Nahorna et al. [(2012). J. Acoust. Soc. Am. 132, 1061-1077] showed that the McGurk effect is reduced by a previous incoherent audiovisual context. This was interpreted as showing the existence of an audiovisual binding stage controlling the fusion process. Incoherence would produce unbinding and decrease the weight of the visual input in fusion. The present paper explores the audiovisual binding system to characterize its dynamics. A first experiment assesses the dynamics of unbinding, and shows that it is rapid: An incoherent context less than 0.5 s long (typically one syllable) suffices to produce a maximal reduction in the McGurk effect. A second experiment tests the rebinding process, by presenting a short period of either coherent material or silence after the incoherent unbinding context. Coherence provides rebinding, with a recovery of the McGurk effect, while silence provides no rebinding and hence freezes the unbinding process. These experiments are interpreted in the framework of an audiovisual speech scene analysis process assessing the perceptual organization of an audiovisual speech input before decision takes place at a higher processing stage. (C) 2015 Acoustical Society of America.
引用
收藏
页码:362 / 377
页数:16
相关论文
共 50 条
  • [1] Binding and unbinding the auditory and visual streams in the McGurk effect
    Nahorna, Olha
    Berthommier, Frederic
    Schwartz, Jean-Luc
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2012, 132 (02) : 1061 - 1077
  • [2] The threshold for the McGurk effect in audio-visual noise decreases with development
    Hirst, Rebecca J.
    Stacey, Jemaine E.
    Cragg, Lucy
    Stacey, Paula C.
    Allen, Harriet A.
    SCIENTIFIC REPORTS, 2018, 8
  • [3] Audio-visual speech experience with age influences perceived audio-visual asynchrony in speech
    Alm, Magnus
    Behne, Dawn
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2013, 134 (04) : 3001 - 3010
  • [4] Audio-Visual Speech Cue Combination
    Arnold, Derek H.
    Tear, Morgan
    Schindel, Ryan
    Roseboom, Warrick
    PLOS ONE, 2010, 5 (04):
  • [5] Do gender differences in audio-visual benefit and visual influence in audio-visual speech perception emerge with age?
    Alm, Magnus
    Behne, Dawn
    FRONTIERS IN PSYCHOLOGY, 2015, 6
  • [6] Somatosensory contribution to audio-visual speech processing
    Ito, Takayuki
    Ohashi, Hiroki
    Gracco, Vincent L.
    CORTEX, 2021, 143 : 195 - 204
  • [7] Effects of aging on audio-visual speech integration Effects of aging on audio-visual speech integration
    Huyse, Aurelie
    Leybaert, Jacqueline
    Berthommier, Frederic
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2014, 136 (04) : 1918 - 1931
  • [8] Animated virtual characters to explore audio-visual speech in controlled and naturalistic environments
    Theze, Raphael
    Gadiri, Mehdi Ali
    Albert, Louis
    Provost, Antoine
    Giraud, Anne-Lise
    Megevand, Pierre
    SCIENTIFIC REPORTS, 2020, 10 (01)
  • [9] Investigating the audio-visual speech detection advantage
    Kim, J
    Davis, C
    SPEECH COMMUNICATION, 2004, 44 (1-4) : 19 - 30
  • [10] Adaptation to Social-Linguistic Associations in Audio-Visual Speech
    Babel, Molly
    BRAIN SCIENCES, 2022, 12 (07)