Transformer-based fusion model for mild depression recognition with EEG and pupil area signals

被引:0
|
作者
Zhu, Jing [1 ]
Li, Yuanlong [1 ]
Yang, Changlin [1 ]
Cai, Hanshu [1 ]
Li, Xiaowei [1 ]
Hu, Bin [1 ,2 ,3 ,4 ,5 ]
机构
[1] Lanzhou Univ, Sch Informat Sci & Engn, Gansu Prov Key Lab Wearable Comp, Lanzhou 73000, Peoples R China
[2] Beijing Inst Technol, Sch Med Technol, Beijing, Peoples R China
[3] Chinese Acad Sci, CAS Ctr Excellence Brain Sci & Intelligence Techno, Shanghai Inst Biol Sci, Shanghai 73000, Peoples R China
[4] Lanzhou Univ, Joint Res Ctr Cognit Neurosensor Technol, Lanzhou, Peoples R China
[5] Chinese Acad Sci, Inst Semicond, Lanzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Mild depression; EEG; Pupil area signal; Transformer; Attention;
D O I
10.1007/s11517-024-03269-8
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Early detection and treatment are crucial for the prevention and treatment of depression; compared with major depression, current researches pay less attention to mild depression. Meanwhile, analysis of multimodal biosignals such as EEG, eye movement data, and magnetic resonance imaging provides reliable technical means for the quantitative analysis of depression. However, how to effectively capture relevant and complementary information between multimodal data so as to achieve efficient and accurate depression recognition remains a challenge. This paper proposes a novel Transformer-based fusion model using EEG and pupil area signals for mild depression recognition. We first introduce CSP into the Transformer to construct single-modal models of EEG and pupil data and then utilize attention bottleneck to construct a mid-fusion model to facilitate information exchange between the two modalities; this strategy enables the model to learn the most relevant and complementary information for each modality and only share the necessary information, which improves the model accuracy while reducing the computational cost. Experimental results show that the accuracy of the EEG and pupil area signals of single-modal models we constructed is 89.75% and 84.17%, the precision is 92.04% and 95.21%, the recall is 89.5% and 71%, the specificity is 90% and 97.33%, the F1 score is 89.41% and 78.44%, respectively, and the accuracy of mid-fusion model can reach 93.25%. Our study demonstrates that the Transformer model can learn the long-term time-dependent relationship between EEG and pupil area signals, providing an idea for designing a reliable multimodal fusion model for mild depression recognition based on EEG and pupil area signals.
引用
收藏
页数:17
相关论文
共 50 条
  • [1] Mutual Information Based Fusion Model (MIBFM): Mild Depression Recognition Using EEG and Pupil Area Signals
    Zhu, Jing
    Yang, Changlin
    Xie, Xiannian
    Wei, Shiqing
    Li, Yizhou
    Li, Xiaowei
    Hu, Bin
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (03) : 2102 - 2115
  • [2] Classification and recognition of gesture EEG signals with Transformer-Based models
    Qu, Yan
    Li, Congsheng
    Jiang, Haoyu
    2024 3RD INTERNATIONAL CONFERENCE ON ROBOTICS, ARTIFICIAL INTELLIGENCE AND INTELLIGENT CONTROL, RAIIC 2024, 2024, : 415 - 418
  • [3] Transformer-based ensemble deep learning model for EEG-based emotion recognition
    Xiaopeng Si
    Dong Huang
    Yulin Sun
    Shudi Huang
    He Huang
    Dong Ming
    Brain Science Advances, 2023, 9 (03) : 210 - 223
  • [4] Transformer-Based Model for Auditory EEG Decoding
    Chen, Jiaxin
    Liu, Yin-Long
    Feng, Rui
    Yuan, Jiahong
    Ling, Zhen-Hua
    MAN-MACHINE SPEECH COMMUNICATION, NCMMSC 2024, 2025, 2312 : 129 - 143
  • [5] ERTNet: an interpretable transformer-based framework for EEG emotion recognition
    Liu, Ruixiang
    Chao, Yihu
    Ma, Xuerui
    Sha, Xianzheng
    Sun, Limin
    Li, Shuo
    Chang, Shijie
    FRONTIERS IN NEUROSCIENCE, 2024, 18
  • [6] A spatial and temporal transformer-based EEG emotion recognition in VR environment
    Li, Ming
    Yu, Peng
    Shen, Yang
    FRONTIERS IN HUMAN NEUROSCIENCE, 2025, 19
  • [7] A transformer-based multi-features fusion model for prediction of conversion in mild cognitive impairment
    Zheng, Guowei
    Zhang, Yu
    Zhao, Ziyang
    Wang, Yin
    Liu, Xia
    Shang, Yingying
    Cong, Zhaoyang
    Dimitriadis, Stavros I.
    Yao, Zhijun
    Hu, Bin
    METHODS, 2022, 204 : 241 - 248
  • [8] A Transformer-Based Fusion Recommendation Model For IPTV Applications
    Li, Heng
    Lei, Hang
    Yang, Maolin
    Zeng, Jinghong
    Zhu, Di
    Fu, Shouwei
    2020 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND BIG DATA (ICAIBD 2020), 2020, : 177 - 182
  • [9] RM-Transformer: A Transformer-based Model for Mandarin Speech Recognition
    Lu, Xingyu
    Hu, Jianguo
    Li, Shenhao
    Ding, Yanyu
    2022 IEEE 2ND INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATION AND ARTIFICIAL INTELLIGENCE (CCAI 2022), 2022, : 194 - 198
  • [10] EEG Classification with Transformer-Based Models
    Sun, Jiayao
    Xie, Jin
    Zhou, Huihui
    2021 IEEE 3RD GLOBAL CONFERENCE ON LIFE SCIENCES AND TECHNOLOGIES (IEEE LIFETECH 2021), 2021, : 92 - 93