The integration of continuous audio and visual speech in a cocktail-party environment depends on attention

被引:11
作者
Ahmed, Farhin [1 ,2 ]
Nidiffer, Aaron R. [1 ,2 ]
O'Sullivan, Aisling E. [1 ,2 ,3 ,4 ]
Zuk, Nathaniel J. [5 ]
Lalor, Edmund C. [1 ,2 ,3 ,4 ]
机构
[1] Univ Rochester, Dept Biomed Engn, Dept Neurosci, Rochester, NY 14627 USA
[2] Univ Rochester, Del Monte Inst Neurosci, Rochester, NY 14627 USA
[3] Trinity Coll Dublin, Trinity Ctr Biomed Engn, Sch Engn, Dublin 2, Ireland
[4] Trinity Coll Dublin, Trinity Coll Inst Neurosci, Dublin 2, Ireland
[5] Hebrew Univ Jerusalem, Edmond & Lily Safra Ctr Brain Sci, Jerusalem, Israel
基金
爱尔兰科学基金会;
关键词
Multisensory integration; Speech; Cocktail party; Hierarchical processing; AUDIOVISUAL SPEECH; MULTISENSORY INTEGRATION; SELECTIVE ATTENTION; AUDITORY-CORTEX; INFORMATION; DIRECTION; TRACKING; OBJECTS; HEAR;
D O I
10.1016/j.neuroimage.2023.120143
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
In noisy environments, our ability to understand speech benefits greatly from seeing the speaker's face. This is attributed to the brain's ability to integrate audio and visual information, a process known as multisensory inte-gration. In addition, selective attention plays an enormous role in what we understand, the so-called cocktail-party phenomenon. But how attention and multisensory integration interact remains incompletely understood, partic-ularly in the case of natural, continuous speech. Here, we addressed this issue by analyzing EEG data recorded from participants who undertook a multisensory cocktail-party task using natural speech. To assess multisensory integration, we modeled the EEG responses to the speech in two ways. The first assumed that audiovisual speech processing is simply a linear combination of audio speech processing and visual speech processing (i.e., an A + V model), while the second allows for the possibility of audiovisual interactions (i.e., an AV model). Applying these models to the data revealed that EEG responses to attended audiovisual speech were better explained by an AV model, providing evidence for multisensory integration. In contrast, unattended audiovisual speech responses were best captured using an A + V model, suggesting that multisensory integration is suppressed for unattended speech. Follow up analyses revealed some limited evidence for early multisensory integration of unattended AV speech, with no integration occurring at later levels of processing. We take these findings as evidence that the integration of natural audio and visual speech occurs at multiple levels of processing in the brain, each of which can be differentially affected by attention.
引用
收藏
页数:13
相关论文
共 69 条
[1]   The CIPICHRTF database [J].
Algazi, VR ;
Duda, RO ;
Thompson, DM ;
Avendano, C .
PROCEEDINGS OF THE 2001 IEEE WORKSHOP ON THE APPLICATIONS OF SIGNAL PROCESSING TO AUDIO AND ACOUSTICS, 2001, :99-102
[2]   Audiovisual integration of speech falters under high attention demands [J].
Alsius, A ;
Navarra, J ;
Campbell, R ;
Soto-Faraco, S .
CURRENT BIOLOGY, 2005, 15 (09) :839-843
[3]   Attention to touch weakens audiovisual speech integration [J].
Alsius, Agnes ;
Navarra, Jordi ;
Soto-Faraco, Salvador .
EXPERIMENTAL BRAIN RESEARCH, 2007, 183 (03) :399-404
[4]   Effect of attentional load on audiovisual speech perception: evidence from ERPs [J].
Alsius, Agnes ;
Moettoenen, Riikka ;
Sams, Mikko E. ;
Soto-Faraco, Salvador ;
Tiippana, Kaisa .
FRONTIERS IN PSYCHOLOGY, 2014, 5
[5]   Integration of Visual Information in Auditory Cortex Promotes Auditory Scene Analysis through Multisensory Binding [J].
Atilgan, Huriye ;
Town, Stephen M. ;
Wood, Katherine C. ;
Jones, Gareth P. ;
Maddox, Ross K. ;
Lee, Adrian K. C. ;
Bizley, Jennifer K. .
NEURON, 2018, 97 (03) :640-+
[6]   Degrading phonetic information affects matching of audiovisual speech in adults, but not in infants [J].
Baart, Martijn ;
Vroomen, Jean ;
Shaw, Kathleen ;
Bortfeld, Heather .
COGNITION, 2014, 130 (01) :31-43
[7]   Integration of auditory and visual information about objects in superior temporal sulcus [J].
Beauchamp, MS ;
Lee, KE ;
Argall, BD ;
Martin, A .
NEURON, 2004, 41 (05) :809-823
[8]   CONTROLLING THE FALSE DISCOVERY RATE - A PRACTICAL AND POWERFUL APPROACH TO MULTIPLE TESTING [J].
BENJAMINI, Y ;
HOCHBERG, Y .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 1995, 57 (01) :289-300
[9]   The ventriloquist effect does not depend on the direction of deliberate visual attention [J].
Bertelson, P ;
Vroomen, J ;
de Gelder, B ;
Driver, J .
PERCEPTION & PSYCHOPHYSICS, 2000, 62 (02) :321-332
[10]   Informational and energetic masking effects in the perception of multiple simultaneous talkers [J].
Brungart, DS ;
Simpson, BD ;
Ericson, MA ;
Scott, KR .
JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2001, 110 (05) :2527-2538