BENDR: Using Transformers and a Contrastive Self-Supervised Learning Task to Learn From Massive Amounts of EEG Data

被引:112
作者
Kostas, Demetres [1 ,2 ]
Aroca-Ouellette, Stephane [1 ,2 ]
Rudzicz, Frank [1 ,2 ,3 ]
机构
[1] Univ Toronto, Dept Comp Sci, Toronto, ON, Canada
[2] Vector Inst Artificial Intelligence, Toronto, ON, Canada
[3] St Michaels Hosp, Li Ka Shing Knowledge Inst, Toronto, ON, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
brain computer interface; deep learning; artificial neural network; transformers; semi-supervised learning; contrastive learning; convolutional neural network; sequence modeling; BRAIN-COMPUTER INTERFACE; CLASSIFICATION;
D O I
10.3389/fnhum.2021.653659
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Deep neural networks (DNNs) used for brain-computer interface (BCI) classification are commonly expected to learn general features when trained across a variety of contexts, such that these features could be fine-tuned to specific contexts. While some success is found in such an approach, we suggest that this interpretation is limited and an alternative would better leverage the newly (publicly) available massive electroencephalography (EEG) datasets. We consider how to adapt techniques and architectures used for language modeling (LM) that appear capable of ingesting awesome amounts of data toward the development of encephalography modeling with DNNs in the same vein. We specifically adapt an approach effectively used for automatic speech recognition, which similarly (to LMs) uses a self-supervised training objective to learn compressed representations of raw data signals. After adaptation to EEG, we find that a single pre-trained model is capable of modeling completely novel raw EEG sequences recorded with differing hardware, and different subjects performing different tasks. Furthermore, both the internal representations of this model and the entire architecture can be fine-tuned to a variety of downstream BCI and EEG classification tasks, outperforming prior work in more task-specific (sleep stage classification) self-supervision.
引用
收藏
页数:15
相关论文
共 76 条
[1]   Performance variation in motor imagery brain-computer interface: A brief review [J].
Ahn, Minkyu ;
Jun, Sung Chan .
JOURNAL OF NEUROSCIENCE METHODS, 2015, 243 :103-110
[2]  
[Anonymous], 2015, ACS SYM SER
[3]  
Aroca-Ouellette S, 2020, PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), P4970
[4]  
Arora Sanjeev, 2019, P MACHINE LEARNING R, V97
[5]  
Ba J., 2020, P MACHINE LEARNING S, P9868
[6]  
Baevski A, 2020, ADV NEUR IN, V33
[7]  
Baevski A, 2020, INT CONF ACOUST SPEE, P7694, DOI [10.1109/icassp40776.2020.9054224, 10.1109/ICASSP40776.2020.9054224]
[8]  
Banville H., 2019 IEEE 29 INT WOR, P1
[9]   Uncovering the structure of clinical EEG signals with self-supervised learning [J].
Banville, Hubert ;
Chehab, Omar ;
Hyvarinen, Aapo ;
Engemann, Denis-Alexander ;
Gramfort, Alexandre .
JOURNAL OF NEURAL ENGINEERING, 2021, 18 (04)
[10]  
Brown TB, 2020, ADV NEUR IN, V33