Decoding the attended speech stream with multi-channel EEG: implications for online, daily-life applications

被引:169
作者
Mirkovic, Bojana [1 ,3 ]
Debener, Stefan [1 ,3 ,4 ]
Jaeger, Manuela [1 ]
De Vos, Maarten [2 ]
机构
[1] Carl von Ossietzky Univ Oldenburg, Dept Psychol, Neuropsychol Lab, D-26129 Oldenburg, Germany
[2] Univ Oxford, Dept Engn, Inst Biomed Engn, Oxford OX3 7DQ, England
[3] Carl von Ossietzky Univ Oldenburg, Cluster Excellence Hearing4all, D-26129 Oldenburg, Germany
[4] Carl von Ossietzky Univ Oldenburg, Res Ctr Neurosensory Sci, D-26129 Oldenburg, Germany
关键词
cocktail party; selective attention; speech envelope; stimulus reconstruction; mobile EEG; BCI; BRAIN-COMPUTER INTERFACE; AUDITORY OBJECTS; RESPONSES; CUES;
D O I
10.1088/1741-2560/12/4/046007
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Objective. Recent studies have provided evidence that temporal envelope driven speech decoding from high-density electroencephalography (EEG) and magnetoencephalography recordings can identify the attended speech stream in a multi-speaker scenario. The present work replicated the previous high density EEG study and investigated the necessary technical requirements for practical attended speech decoding with EEG. Approach. Twelve normal hearing participants attended to one out of two simultaneously presented audiobook stories, while high density EEG was recorded. An offline iterative procedure eliminating those channels contributing the least to decoding provided insight into the necessary channel number and optimal cross-subject channel configuration. Aiming towards the future goal of near real-time classification with an individually trained decoder, the minimum duration of training data necessary for successful classification was determined by using a chronological cross-validation approach. Main results. Close replication of the previously reported results confirmed the method robustness. Decoder performance remained stable from 96 channels down to 25. Furthermore, for less than 15 min of training data, the subject-independent (pre-trained) decoder performed better than an individually trained decoder did. Significance. Our study complements previous research and provides information suggesting that efficient low-density EEG online decoding is within reach.
引用
收藏
页数:9
相关论文
共 37 条
[1]   Human cortical responses to the speech envelope [J].
Aiken, Steven J. ;
Picton, Terence W. .
EAR AND HEARING, 2008, 29 (02) :139-157
[2]   The non-invasive Berlin Brain-Computer Interface:: Fast acquisition of effective performance in untrained subjects [J].
Blankertz, Benjamin ;
Dornhege, Guido ;
Krauledat, Matthias ;
Mueller, Klaus-Robert ;
Curio, Gabriel .
NEUROIMAGE, 2007, 37 (02) :539-550
[3]   Denoising based on spatial filtering [J].
de Cheveigne, Alain ;
Simon, Jonathan Z. .
JOURNAL OF NEUROSCIENCE METHODS, 2008, 171 (02) :331-339
[4]   P300 speller BCI with a mobile EEG system: comparison to a traditional amplifier [J].
De Vos, Maarten ;
Kroesen, Markus ;
Emkes, Reiner ;
Debener, Stefan .
JOURNAL OF NEURAL ENGINEERING, 2014, 11 (03)
[5]   Towards a truly mobile auditory brain-computer interface: Exploring the P300 to take away [J].
De Vos, Maarten ;
Gandras, Katharina ;
Debener, Stefan .
INTERNATIONAL JOURNAL OF PSYCHOPHYSIOLOGY, 2014, 91 (01) :46-53
[6]   Mobile EEG: Towards brain activity monitoring during natural action and cognition [J].
De Vos, Maarten ;
Debener, Stefan .
INTERNATIONAL JOURNAL OF PSYCHOPHYSIOLOGY, 2014, 91 (01) :1-2
[7]   How about taking a low-cost, small, and wireless EEG for a walk? [J].
Debener, Stefan ;
Minow, Falk ;
Emkes, Reiner ;
Gandras, Katharina ;
De Vos, Maarten .
PSYCHOPHYSIOLOGY, 2012, 49 (11) :1617-1621
[8]   EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysis [J].
Delorme, A ;
Makeig, S .
JOURNAL OF NEUROSCIENCE METHODS, 2004, 134 (01) :9-21
[9]   Cortical entrainment to continuous speech: functional roles and interpretations [J].
Ding, Nai ;
Simon, Jonathan Z. .
FRONTIERS IN HUMAN NEUROSCIENCE, 2014, 8
[10]   Emergence of neural encoding of auditory objects while listening to competing speakers [J].
Ding, Nai ;
Simon, Jonathan Z. .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2012, 109 (29) :11854-11859