Auditory attention tracking states in a cocktail party environment can be decoded by deep convolutional neural networks

被引:13
作者
Tian, Yin [1 ]
Ma, Liang [1 ]
机构
[1] ChongQing Univ Posts & Telecommun, Bioinformat Coll, Chongqing 400065, Peoples R China
基金
中国国家自然科学基金;
关键词
auditory attention tracking; electroencephalogram (EEG); convolutional neural network (CNN); brain-computer interface (BCI); deep learning (DL); ALPHA OSCILLATIONS; BAND OSCILLATIONS; EEG; SPEECH; BETA; FEEDFORWARD; FEEDBACK; CORTEX; TIME;
D O I
10.1088/1741-2552/ab92b2
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Objective.A deep convolutional neural network (CNN) is a method for deep learning (DL). It has a powerful ability to automatically extract features and is widely used in classification tasks with scalp electroencephalogram (EEG) signals. However, the small number of samples and low signal-to-noise ratio involved in scalp EEG with low spatial resolution constitute a limitation that might restrict potential brain-computer interface (BCI) applications that are based on the CNN model. In the present study, a novel CNN model with source-spatial feature images (SSFIs) as the input is proposed to decode auditory attention tracking states in a cocktail party environment.Approach.We first extract SSFIs using rhythm entropy and weighted minimum norm estimation. Next, we develop a CNN model with three convolutional layers. Furthermore, we estimate the performance of the proposed model via generalized performance, alternative models that deleted or replaced a model's component, and loss curves. Finally, we use a deep transfer model with fine-tuning for a low (poor) behavioral performance group (L-group).Main results.Based on cortical activity reconstructions from the scalp EEGs, the classification accuracy (CA) of the proposed model is 80.4% (chance level: 52.5%), which is superior to that achieved by scalp EEG. Additionally, the performance of the proposed model is more stable when compared to alternative models that delete or replace specific model components. The proposed model identifies the difference between two auditory attention tracking states (successful versus unsuccessful) at an early stage with a short time window (250 ms after target offset). Furthermore, we propose a deep transfer learning model to improve the classification for the L-group. With this model, the CA of the L-group significantly increase by 5.3%.Significance.Our proposed model improves the performance of a decoder for auditory attention tracking, which could be suitable for relieving the difficulty with the attentional modulation of individual's neural responses. It provides a novel communication channel with auditory cognitive BCI for patients with attention and hearing impairment.
引用
收藏
页数:17
相关论文
共 81 条
[1]  
[Anonymous], 2007, ENCY MEASUREMENT STA
[2]  
[Anonymous], 2016, ARXIV160808851
[3]  
[Anonymous], P 14 INT C ART INT S
[4]   A Decade of EEG Theta/Beta Ratio Research in ADHD: A Meta-Analysis [J].
Arns, Martijn ;
Conners, C. Keith ;
Kraemer, Helena C. .
JOURNAL OF ATTENTION DISORDERS, 2013, 17 (05) :374-383
[5]   Alpha Synchrony and the Neurofeedback Control of Spatial Attention [J].
Bagherzadeh, Yasaman ;
Baldauf, Daniel ;
Pantazis, Dimitrios ;
Desimone, Robert .
NEURON, 2020, 105 (03) :577-+
[6]  
Bashivan P., 2015, Learning representations from EEG with deep recurrent-convolutional neural networks
[7]   Visual Areas Exert Feedforward and Feedback Influences through Distinct Frequency Channels [J].
Bastos, Andre Moraes ;
Vezoli, Julien ;
Bosman, Conrado Arturo ;
Schoffelen, Jan-Mathijs ;
Oostenveld, Robert ;
Dowdall, Jarrod Robert ;
De Weerd, Peter ;
Kennedy, Henry ;
Fries, Pascal .
NEURON, 2015, 85 (02) :390-401
[8]  
Chen KX, 2019, AAAI CONF ARTIF INTE, P3321
[10]   Comparison of Two-Talker Attention Decoding from EEG with Nonlinear Neural Networks and Linear Methods [J].
Ciccarelli, Gregory ;
Nolan, Michael ;
Perricone, Joseph ;
Calamia, Paul T. ;
Haro, Stephanie ;
O'Sullivan, James ;
Mesgarani, Nima ;
Quatieri, Thomas F. ;
Smalt, Christopher J. .
SCIENTIFIC REPORTS, 2019, 9 (1)