EEG-based auditory attention decoding with audiovisual speech for hearing-impaired listeners

被引:5
作者
Wang, Bo [1 ]
Xu, Xiran [1 ]
Niu, Yadong [1 ]
Wu, Chao [2 ]
Wu, Xihong [1 ,3 ]
Chen, Jing [1 ,3 ,4 ]
机构
[1] Peking Univ, Speech & Hearing Res Ctr, Sch Intelligence Sci & Technol, Key Lab Machine Percept,Minist Educ, Beijing 100871, Peoples R China
[2] Peking Univ, Sch Nursing, Beijing 100191, Peoples R China
[3] Coll Future Technol, Natl Biomed Imaging Ctr, Beijing 100871, Peoples R China
[4] Room 2227, Sci Bldg 2, 5 Yiheyuan Rd, Beijing 100871, Peoples R China
关键词
audiovisual speech; auditory attention decoding; EEG; hearing impairment; speech-in-noise; TEMPORAL-FINE-STRUCTURE; TO-NOISE RATIOS; OLDER-ADULTS; CORTICAL REPRESENTATION; MODULATION DETECTION; GAP DETECTION; AID USERS; TRACKING; CUES; ENTRAINMENT;
D O I
10.1093/cercor/bhad325
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Auditory attention decoding (AAD) was used to determine the attended speaker during an auditory selective attention task. However, the auditory factors modulating AAD remained unclear for hearing-impaired (HI) listeners. In this study, scalp electroencephalogram (EEG) was recorded with an auditory selective attention paradigm, in which HI listeners were instructed to attend one of the two simultaneous speech streams with or without congruent visual input (articulation movements), and at a high or low target-to-masker ratio (TMR). Meanwhile, behavioral hearing tests (i.e. audiogram, speech reception threshold, temporal modulation transfer function) were used to assess listeners' individual auditory abilities. The results showed that both visual input and increasing TMR could significantly enhance the cortical tracking of the attended speech and AAD accuracy. Further analysis revealed that the audiovisual (AV) gain in attended speech cortical tracking was significantly correlated with listeners' auditory amplitude modulation (AM) sensitivity, and the TMR gain in attended speech cortical tracking was significantly correlated with listeners' hearing thresholds. Temporal response function analysis revealed that subjects with higher AM sensitivity demonstrated more AV gain over the right occipitotemporal and bilateral frontocentral scalp electrodes.
引用
收藏
页码:10972 / 10983
页数:12
相关论文
共 88 条
[1]   Assessing variability in audiovisual speech integration skills using capacity and accuracy measures [J].
Altieri, Nicholas ;
Hudock, Daniel .
INTERNATIONAL JOURNAL OF AUDIOLOGY, 2014, 53 (10) :710-718
[2]   Perception of temporal fine-structure cues in speech with minimal envelope cues for listeners with mild-to-moderate hearing loss [J].
Ardoint, Marine ;
Sheft, Stanley ;
Fleuriot, Pierre ;
Garnier, Stephane ;
Lorenzi, Christian .
INTERNATIONAL JOURNAL OF AUDIOLOGY, 2010, 49 (11) :823-831
[3]   Cognitive-Driven Binaural Beamforming Using EEG-Based Auditory Attention Decoding [J].
Aroudi, Ali ;
Doclo, Simon .
IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2020, 28 :862-875
[4]   CONTROLLING THE FALSE DISCOVERY RATE - A PRACTICAL AND POWERFUL APPROACH TO MULTIPLE TESTING [J].
BENJAMINI, Y ;
HOCHBERG, Y .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 1995, 57 (01) :289-300
[5]   SEGREGATION OF GLOBAL AND LOCAL MOTION PROCESSING IN PRIMATE MIDDLE TEMPORAL VISUAL AREA [J].
BORN, RT ;
TOOTELL, RBH .
NATURE, 1992, 357 (6378) :497-499
[6]   The processing of audio-visual speech: empirical and neural bases [J].
Campbell, Ruth .
PHILOSOPHICAL TRANSACTIONS OF THE ROYAL SOCIETY B-BIOLOGICAL SCIENCES, 2008, 363 (1493) :1001-1010
[7]   The Natural Statistics of Audiovisual Speech [J].
Chandrasekaran, Chandramouli ;
Trubanova, Andrea ;
Stillittano, Sebastien ;
Caplier, Alice ;
Ghazanfar, Asif A. .
PLOS COMPUTATIONAL BIOLOGY, 2009, 5 (07)
[8]   Attempt to Predict Temporal Modulation Transfer Function by Amplitude Modulation Following Responses [J].
Chen, Jing ;
Fu, Zhen ;
Wu, Jiping ;
Wu, Xihong .
ACTA ACUSTICA UNITED WITH ACUSTICA, 2018, 104 (05) :821-824
[10]  
Clark J G, 1981, ASHA, V23, P493