Transformer-based fusion model for mild depression recognition with EEG and pupil area signals

被引:0
|
作者
Zhu, Jing [1 ]
Li, Yuanlong [1 ]
Yang, Changlin [1 ]
Cai, Hanshu [1 ]
Li, Xiaowei [1 ]
Hu, Bin [1 ,2 ,3 ,4 ,5 ]
机构
[1] Lanzhou Univ, Sch Informat Sci & Engn, Gansu Prov Key Lab Wearable Comp, Lanzhou 73000, Peoples R China
[2] Beijing Inst Technol, Sch Med Technol, Beijing, Peoples R China
[3] Chinese Acad Sci, CAS Ctr Excellence Brain Sci & Intelligence Techno, Shanghai Inst Biol Sci, Shanghai 73000, Peoples R China
[4] Lanzhou Univ, Joint Res Ctr Cognit Neurosensor Technol, Lanzhou, Peoples R China
[5] Chinese Acad Sci, Inst Semicond, Lanzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Mild depression; EEG; Pupil area signal; Transformer; Attention;
D O I
10.1007/s11517-024-03269-8
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Early detection and treatment are crucial for the prevention and treatment of depression; compared with major depression, current researches pay less attention to mild depression. Meanwhile, analysis of multimodal biosignals such as EEG, eye movement data, and magnetic resonance imaging provides reliable technical means for the quantitative analysis of depression. However, how to effectively capture relevant and complementary information between multimodal data so as to achieve efficient and accurate depression recognition remains a challenge. This paper proposes a novel Transformer-based fusion model using EEG and pupil area signals for mild depression recognition. We first introduce CSP into the Transformer to construct single-modal models of EEG and pupil data and then utilize attention bottleneck to construct a mid-fusion model to facilitate information exchange between the two modalities; this strategy enables the model to learn the most relevant and complementary information for each modality and only share the necessary information, which improves the model accuracy while reducing the computational cost. Experimental results show that the accuracy of the EEG and pupil area signals of single-modal models we constructed is 89.75% and 84.17%, the precision is 92.04% and 95.21%, the recall is 89.5% and 71%, the specificity is 90% and 97.33%, the F1 score is 89.41% and 78.44%, respectively, and the accuracy of mid-fusion model can reach 93.25%. Our study demonstrates that the Transformer model can learn the long-term time-dependent relationship between EEG and pupil area signals, providing an idea for designing a reliable multimodal fusion model for mild depression recognition based on EEG and pupil area signals.
引用
收藏
页数:17
相关论文
共 50 条
  • [21] A Transformer-Based Math Language Model for Handwritten Math Expression Recognition
    Huy Quang Ung
    Cuong Tuan Nguyen
    Hung Tuan Nguyen
    Thanh-Nghia Truong
    Nakagawa, Masaki
    DOCUMENT ANALYSIS AND RECOGNITION, ICDAR 2021, PT II, 2021, 12917 : 403 - 415
  • [22] Temporal fusion transformer-based prediction in aquaponics
    Metin, Ahmet
    Kasif, Ahmet
    Catal, Cagatay
    JOURNAL OF SUPERCOMPUTING, 2023, 79 (17): : 19934 - 19958
  • [23] TDFNet: Transformer-Based Deep-Scale Fusion Network for Multimodal Emotion Recognition
    Zhao, Zhengdao
    Wang, Yuhua
    Shen, Guang
    Xu, Yuezhu
    Zhang, Jiayuan
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 3771 - 3782
  • [24] Enhanced Chinese Named Entity Recognition with Transformer-Based Multi-feature Fusion
    Zhang, Xiaoli
    Zhang, Quan
    Liang, Kun
    Wang, Haoyu
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT III, ICIC 2024, 2024, 14864 : 132 - 141
  • [25] Transformer-Based Multilingual Speech Emotion Recognition Using Data Augmentation and Feature Fusion
    Al-onazi, Badriyya B.
    Nauman, Muhammad Asif
    Jahangir, Rashid
    Malik, Muhmmad Mohsin
    Alkhammash, Eman H.
    Elshewey, Ahmed M.
    APPLIED SCIENCES-BASEL, 2022, 12 (18):
  • [26] Temporal fusion transformer-based prediction in aquaponics
    Ahmet Metin
    Ahmet Kasif
    Cagatay Catal
    The Journal of Supercomputing, 2023, 79 : 19934 - 19958
  • [27] Transformer-based Dynamic Fusion Clustering Network
    Zhang, Chunchun
    Zhao, Yaliang
    Wang, Jinke
    KNOWLEDGE-BASED SYSTEMS, 2022, 258
  • [28] Adaptation of Transformer-Based Models for Depression Detection
    Adebanji, Olaronke O.
    Ojo, Olumide E.
    Calvo, Hiram
    Gelbukh, Irina
    Sidorov, Grigori
    COMPUTACION Y SISTEMAS, 2024, 28 (01): : 151 - 165
  • [29] TIRec: Transformer-based Invoice Text Recognition
    Chen, Yanlan
    2023 2ND ASIA CONFERENCE ON ALGORITHMS, COMPUTING AND MACHINE LEARNING, CACML 2023, 2023, : 175 - 180
  • [30] Transformer-Based Turkish Automatic Speech Recognition
    Tasar, Davut Emre
    Koruyan, Kutan
    Cilgin, Cihan
    ACTA INFOLOGICA, 2024, 8 (01): : 1 - 10