Learning a robust unified domain adaptation framework for cross-subject EEG-based emotion recognition

被引:16
|
作者
Jimenez-Guarneros, Magdiel [1 ]
Fuentes-Pineda, Gibran [1 ]
机构
[1] Univ Nacl Autonoma Mexico, Dept Comp Sci, Inst Invest Matemat Aplicadas & Sistemas, Circuito Escolar S-N,Ciudad Univ, Mexico City 04510, Mexico
关键词
Unsupervised domain adaptation; Deep learning; Emotion recognition; Electroencephalogram; NEURAL-NETWORKS;
D O I
10.1016/j.bspc.2023.105138
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Over the last few years, unsupervised domain adaptation (UDA) based on deep learning has emerged as a solution to build cross-subject emotion recognition models from Electroencephalogram (EEG) signals, aligning the subject distributions within a latent feature space. However, most reported works have a common intrinsic limitation: the subject distribution alignment is coarse-grained, but not all of the feature space is shared between subjects. In this paper, we propose a robust unified domain adaptation framework, named Multi-source Feature Alignment and Label Rectification (MFA-LR), which performs a fine-grained domain alignment at subject and class levels, while inter-class separation and robustness against input perturbations are encouraged in coarse grain. As a complementary step, a pseudo-labeling correction procedure is used to rectify mislabeled target samples. Our proposal was assessed over two public datasets, SEED and SEED-IV, on each of the three available sessions, using leave-one-subject-out cross-validation. Experimental results show an accuracy performance of up to 89.11 & PLUSMN; 07.72% and 74.99 & PLUSMN; 12.10% for the best session on SEED and SEED-IV, as well as an average accuracy of 85.27% and 69.58% on all three sessions, outperforming state-of-the-art results.
引用
收藏
页数:13
相关论文
共 50 条
  • [21] Multisource Transfer Learning for Cross-Subject EEG Emotion Recognition
    Li, Jinpeng
    Qiu, Shuang
    Shen, Yuan-Yuan
    Liu, Cheng-Lin
    He, Huiguang
    IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (07) : 3281 - 3293
  • [22] Cross-Subject Channel Selection Using Modified Relief and Simplified CNN-Based Deep Learning for EEG-Based Emotion Recognition
    Farokhah, Lia
    Sarno, Riyanarto
    Fatichah, Chastine
    IEEE ACCESS, 2023, 11 : 110136 - 110150
  • [23] Personal-Zscore: Eliminating Individual Difference for EEG-Based Cross-Subject Emotion Recognition
    Chen, Huayu
    Sun, Shuting
    Li, Jianxiu
    Yu, Ruilan
    Li, Nan
    Li, Xiaowei
    Hu, Bin
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (03) : 2077 - 2088
  • [24] Interpretable Cross-Subject EEG-Based Emotion Recognition Using Channel-Wise Features†
    Jin, Longbin
    Kim, Eun Yi
    SENSORS, 2020, 20 (23) : 1 - 18
  • [25] WGAN Domain Adaptation for EEG-Based Emotion Recognition
    Luo, Yun
    Zhang, Si-Yang
    Zheng, Wei-Long
    Lu, Bao-Liang
    NEURAL INFORMATION PROCESSING (ICONIP 2018), PT V, 2018, 11305 : 275 - 286
  • [26] Multi-source Selective Graph Domain Adaptation Network for cross-subject EEG emotion recognition
    Wang, Jing
    Ning, Xiaojun
    Xu, Wei
    Li, Yunze
    Jia, Ziyu
    Lin, Youfang
    NEURAL NETWORKS, 2024, 180
  • [27] Dynamic Domain Adaptation for Class-Aware Cross-Subject and Cross-Session EEG Emotion Recognition
    Li, Zhunan
    Zhu, Enwei
    Jin, Ming
    Fan, Cunhang
    He, Huiguang
    Cai, Ting
    Li, Jinpeng
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2022, 26 (12) : 5964 - 5973
  • [28] Gusa: Graph-Based Unsupervised Subdomain Adaptation for Cross-Subject EEG Emotion Recognition
    Li, Xiaojun
    Chen, C. L. Philip
    Chen, Bianna
    Zhang, Tong
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (03) : 1451 - 1462
  • [29] FMLAN: A novel framework for cross-subject and cross-session EEG emotion recognition
    Yu, Peng
    He, Xiaopeng
    Li, Haoyu
    Dou, Haowen
    Tan, Yeyu
    Wu, Hao
    Chen, Badong
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2025, 100
  • [30] CFDA-CSF: A Multi-Modal Domain Adaptation Method for Cross-Subject Emotion Recognition
    Jimenez-Guarneros, Magdiel
    Fuentes-Pineda, Gibran
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (03) : 1502 - 1513