An empirical comparison of deep learning explainability approaches for EEG using simulated ground truth

被引:17
作者
Ravindran, Akshay Sujatha [1 ,2 ,3 ]
Contreras-Vidal, Jose [1 ,2 ]
机构
[1] Univ Houston, Dept Elect & Comp Engn, Noninvas Brain Machine Interface Syst Lab, Houston, TX 77204 USA
[2] Univ Houston, IUCRC BRAIN, Houston, TX 77204 USA
[3] Alto Neurosci, Los Altos, CA 94022 USA
关键词
BRAIN-COMPUTER INTERFACES; NEURAL-NETWORKS; MACHINE INTERFACES; ACTIVATION; MOTOR;
D O I
10.1038/s41598-023-43871-8
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Recent advancements in machine learning and deep learning (DL) based neural decoders have significantly improved decoding capabilities using scalp electroencephalography (EEG). However, the interpretability of DL models remains an under-explored area. In this study, we compared multiple model explanation methods to identify the most suitable method for EEG and understand when some of these approaches might fail. A simulation framework was developed to evaluate the robustness and sensitivity of twelve back-propagation-based visualization methods by comparing to ground truth features. Multiple methods tested here showed reliability issues after randomizing either model weights or labels: e.g., the saliency approach, which is the most used visualization technique in EEG, was not class or model-specific. We found that DeepLift was consistently accurate as well as robust to detect the three key attributes tested here (temporal, spatial, and spectral precision). Overall, this study provides a review of model explanation methods for DL-based neural decoders and recommendations to understand when some of these methods fail and what they can capture in EEG.
引用
收藏
页数:20
相关论文
共 80 条
[1]  
Adebayo J, 2020, Arxiv, DOI arXiv:1810.03292
[2]   Convolutional neural networks for decoding electroencephalography responses and visualizing trial by trial changes in discriminant features [J].
Aellen, Florence M. ;
Goktepe-Kavis, Pinar ;
Apostolopoulos, Stefanos ;
Tzovara, Athina .
JOURNAL OF NEUROSCIENCE METHODS, 2021, 364
[3]   Deep learning for motor imagery EEG-based classification: A review [J].
Al-Saegh, Ali ;
Dawwd, Shefa A. ;
Abdul-Jabbar, Jassim M. .
BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2021, 63
[4]  
Ancona M, 2018, Arxiv, DOI arXiv:1711.06104
[5]  
Arras L, 2021, Arxiv, DOI arXiv:2003.07258
[6]   On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation [J].
Bach, Sebastian ;
Binder, Alexander ;
Montavon, Gregoire ;
Klauschen, Frederick ;
Mueller, Klaus-Robert ;
Samek, Wojciech .
PLOS ONE, 2015, 10 (07)
[7]  
Baehrens D, 2010, J MACH LEARN RES, V11, P1803
[8]   A Lightweight Multi-Scale Convolutional Neural Network for P300 Decoding: Analysis of Training Strategies and Uncovering of Network Decision [J].
Borra, Davide ;
Fantozzi, Silvia ;
Magosso, Elisa .
FRONTIERS IN HUMAN NEUROSCIENCE, 2021, 15
[9]   On the Mathematical Properties of the Structural Similarity Index [J].
Brunet, Dominique ;
Vrscay, Edward R. ;
Wang, Zhou .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2012, 21 (04) :1488-1499
[10]   Understanding adversarial examples requires a theory of artefacts for deep learning [J].
Buckner, Cameron .
NATURE MACHINE INTELLIGENCE, 2020, 2 (12) :731-736