Emotion recognition using hierarchical spatial-temporal learning transformer from regional to global brain

被引:4
|
作者
Cheng, Cheng [1 ]
Liu, Wenzhe [2 ]
Feng, Lin [1 ,3 ]
Jia, Ziyu [4 ]
机构
[1] Dalian Univ Technol, Dept Comp Sci & Technol, Dalian, Peoples R China
[2] Huzhou Univ, Sch Informat Engn, Huzhou, Peoples R China
[3] Dalian Minzu Univ, Sch Informat & Commun Engn, Dlian, Peoples R China
[4] Univ Chinese Acad Sci, Chinese Acad Sci, Brainnetome Ctr, Inst Automat, Beijing, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Emotion recognition; Electroencephalogram (EEG); Transformer; Spatiotemporal features; EEG; FUSION;
D O I
10.1016/j.neunet.2024.106624
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Emotion recognition is an essential but challenging task in human-computer interaction systems due to the distinctive spatial structures and dynamic temporal dependencies associated with each emotion. However, current approaches fail to accurately capture the intricate effects of electroencephalogram (EEG) signals across different brain regions on emotion recognition. Therefore, this paper designs a transformer-based method, denoted by R2G-STLT, which relies on a spatial-temporal transformer encoder with regional to global hierarchical learning that learns the representative spatiotemporal features from the electrode level to the brain-region level. The regional spatial-temporal transformer (RST-Trans) encoder is designed to obtain spatial information and context dependence at the electrode level aiming to learn the regional spatiotemporal features. Then, the global spatial-temporal transformer (GST-Trans) encoder is utilized to extract reliable global spatiotemporal features, reflecting the impact of various brain regions on emotion recognition tasks. Moreover, the multi-head attention mechanism is placed into the GST-Trans encoder to empower it to capture the longrange spatial-temporal information among the brain regions. Finally, subject-independent experiments are conducted on each frequency band of the DEAP, SEED, and SEED-IV datasets to assess the performance of the proposed model. Results indicate that the R2G-STLT model surpasses several state-of-the-art approaches.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] ASTT: acoustic spatial-temporal transformer for short utterance speaker recognition
    Xing Wu
    Ruixuan Li
    Bin Deng
    Ming Zhao
    Xingyue Du
    Jianjia Wang
    Kai Ding
    Multimedia Tools and Applications, 2023, 82 : 33039 - 33061
  • [22] ASTT: acoustic spatial-temporal transformer for short utterance speaker recognition
    Wu, Xing
    Li, Ruixuan
    Deng, Bin
    Zhao, Ming
    Du, Xingyue
    Wang, Jianjia
    Ding, Kai
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (21) : 33039 - 33061
  • [23] Transformer-Based Multimodal Spatial-Temporal Fusion for Gait Recognition
    Zhang, Jikai
    Ji, Mengyu
    He, Yihao
    Guo, Dongliang
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT XV, 2025, 15045 : 494 - 507
  • [24] GC-STCL: A Granger Causality-Based Spatial-Temporal Contrastive Learning Framework for EEG Emotion Recognition
    Wang, Lei
    Wang, Siming
    Jin, Bo
    Wei, Xiaopeng
    ENTROPY, 2024, 26 (07)
  • [25] Spatial-temporal transformer for end-to-end sign language recognition
    Cui, Zhenchao
    Zhang, Wenbo
    Li, Zhaoxin
    Wang, Zhaoqi
    COMPLEX & INTELLIGENT SYSTEMS, 2023, 9 (04) : 4645 - 4656
  • [26] RESTHT: relation-enhanced spatial-temporal hierarchical transformer for video captioning
    Zheng, Lihuan
    Xu, Wanru
    Miao, Zhenjiang
    Qiu, Xinxiu
    Gong, Shanshan
    VISUAL COMPUTER, 2025, 41 (01) : 591 - 604
  • [27] Subject-independent emotion recognition of EEG signals using graph attention-based spatial-temporal pattern learning
    Zhu, Yiwen
    Guo, Yeshuang
    Zhu, Wenzhe
    Di, Lare
    Yin, Thong
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 7070 - 7075
  • [28] Learning Complementary Spatial-Temporal Transformer for Video Salient Object Detection
    Liu, Nian
    Nan, Kepan
    Zhao, Wangbo
    Yao, Xiwen
    Han, Junwei
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (08) : 10663 - 10673
  • [29] Pyramid Spatial-Temporal Graph Transformer for Skeleton-Based Action Recognition
    Chen, Shuo
    Xu, Ke
    Jiang, Xinghao
    Sun, Tanfeng
    APPLIED SCIENCES-BASEL, 2022, 12 (18):
  • [30] STST: Spatial-Temporal Specialized Transformer for Skeleton-based Action Recognition
    Zhang, Yuhan
    Wu, Bo
    Li, Wen
    Duan, Lixin
    Gan, Chuang
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 3229 - 3237