Audio-visual spatial alignment improves integration in the presence of a competing audio-visual stimulus

被引:15
|
作者
Fleming, Justin T. [1 ]
Noyce, Abigail L. [2 ]
Shinn-Cunningham, Barbara G. [3 ]
机构
[1] Harvard Med Sch, Div Med Sci, Speech & Hearing Biosci & Technol Program, Boston, MA USA
[2] Boston Univ, Dept Psychol & Brain Sci, Boston, MA 02215 USA
[3] Carnegie Mellon Univ, Neurosci Inst, Pittsburgh, PA 15213 USA
关键词
Audio-visual integration; Attention; Visual search; Electroencephalography; Temporal coherence; Spatial alignment; AUDITORY-VISUAL INTERACTIONS; EVENT-RELATED POTENTIALS; TEMPORAL BINDING WINDOW; MULTISENSORY INTEGRATION; SELECTIVE ATTENTION; INDIVIDUAL-DIFFERENCES; REACTION-TIME; CORTEX; HUMANS; RESPONSES;
D O I
10.1016/j.neuropsychologia.2020.107530
中图分类号
B84 [心理学]; C [社会科学总论]; Q98 [人类学];
学科分类号
03 ; 0303 ; 030303 ; 04 ; 0402 ;
摘要
In order to parse the world around us, we must constantly determine which sensory inputs arise from the same physical source and should therefore be perceptually integrated. Temporal coherence between auditory and visual stimuli drives audio-visual (AV) integration, but the role played by AV spatial alignment is less well understood. Here, we manipulated AV spatial alignment and collected electroencephalography (EEG) data while human subjects performed a free-field variant of the "pip and pop" AV search task. In this paradigm, visual search is aided by a spatially uninformative auditory tone, the onsets of which are synchronized to changes in the visual target. In Experiment 1, tones were either spatially aligned or spatially misaligned with the visual display. Regardless of AV spatial alignment, we replicated the key pip and pop result of improved AV search times. Mirroring the behavioral results, we found an enhancement of early event-related potentials (ERPs), particularly the auditory N1 component, in both AV conditions. We demonstrate that both top-down and bottom-up attention contribute to these N1 enhancements. In Experiment 2, we tested whether spatial alignment influences AV integration in a more challenging context with competing multisensory stimuli. An AV foil was added that visually resembled the target and was synchronized to its own stream of synchronous tones. The visual components of the AV target and AV foil occurred in opposite hemifields; the two auditory components were also in opposite hemifields and were either spatially aligned or spatially misaligned with the visual components to which they were synchronized. Search was fastest when the auditory and visual components of the AV target (and the foil) were spatially aligned. Attention modulated ERPs in both spatial conditions, but importantly, the scalp topography of early evoked responses shifted only when stimulus components were spatially aligned, signaling the recruitment of different neural generators likely related to multisensory integration. These results suggest that AV integration depends on AV spatial alignment when stimuli in both modalities compete for selective integration, a common scenario in real-world perception.
引用
收藏
页数:17
相关论文
共 50 条
  • [41] Effects of stimulus intensity on audio-visual integration: evidence from ERPs
    Li, Zimo
    Yang, Weiping
    Li, Shengnan
    Guo, Ao
    INTERNATIONAL JOURNAL OF PSYCHOLOGY, 2023, 58 : 222 - 222
  • [42] An audio-visual speech recognition system for testing new audio-visual databases
    Pao, Tsang-Long
    Liao, Wen-Yuan
    VISAPP 2006: PROCEEDINGS OF THE FIRST INTERNATIONAL CONFERENCE ON COMPUTER VISION THEORY AND APPLICATIONS, VOL 2, 2006, : 192 - +
  • [43] Audio-visual event detection based on mining of semantic audio-visual labels
    Goh, KS
    Miyahara, K
    Radhakrishan, R
    Xiong, ZY
    Divakaran, A
    STORAGE AND RETRIEVAL METHODS AND APPLICATIONS FOR MULTIMEDIA 2004, 2004, 5307 : 292 - 299
  • [44] LEARNING CONTEXTUALLY FUSED AUDIO-VISUAL REPRESENTATIONS FOR AUDIO-VISUAL SPEECH RECOGNITION
    Zhang, Zi-Qiang
    Zhang, Jie
    Zhang, Jian-Shu
    Wu, Ming-Hui
    Fang, Xin
    Dai, Li-Rong
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 1346 - 1350
  • [45] AV2AV: Direct Audio-Visual Speech to Audio-Visual Speech Translation with Unified Audio-Visual Speech Representation
    Choi, Jeongsoo
    Park, Se Jin
    Kim, Minsu
    Ro, Yong Man
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 27315 - 27327
  • [46] The Development of Audio-Visual Integration for Temporal Judgements
    Adams, Wendy J.
    PLOS COMPUTATIONAL BIOLOGY, 2016, 12 (04)
  • [47] Audio-visual integration of emotional cues in song
    Thompson, William Forde
    Russo, Frank A.
    Quinto, Lena
    COGNITION & EMOTION, 2008, 22 (08) : 1457 - 1470
  • [48] THE ROLE OF INSULA IN PROCESSES OF AUDIO-VISUAL INTEGRATION
    Kostanyan, Daria
    Vartanov, Alexander
    Kiselnikov, Andrey
    Kozlovskiy, Stanislav
    Ushakov, Vadim
    Buldakova, Natalia
    Korosteleva, Anastasiya
    PSYCHOPHYSIOLOGY, 2018, 55 : S133 - S133
  • [49] Audio-visual integration in the perception of tap dancing
    Arrighi, R.
    Marini, F.
    Burr, D.
    PERCEPTION, 2007, 36 : 172 - 172
  • [50] Audio-visual sensory integration and mismatch negativity
    Ullsperger, P
    Erdmann, U
    Freude, G
    Dehoff, W
    INTERNATIONAL JOURNAL OF PSYCHOPHYSIOLOGY, 2002, 45 (1-2) : 127 - 127