Emotion recognition using hierarchical spatial-temporal learning transformer from regional to global brain

被引:4
|
作者
Cheng, Cheng [1 ]
Liu, Wenzhe [2 ]
Feng, Lin [1 ,3 ]
Jia, Ziyu [4 ]
机构
[1] Dalian Univ Technol, Dept Comp Sci & Technol, Dalian, Peoples R China
[2] Huzhou Univ, Sch Informat Engn, Huzhou, Peoples R China
[3] Dalian Minzu Univ, Sch Informat & Commun Engn, Dlian, Peoples R China
[4] Univ Chinese Acad Sci, Chinese Acad Sci, Brainnetome Ctr, Inst Automat, Beijing, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Emotion recognition; Electroencephalogram (EEG); Transformer; Spatiotemporal features; EEG; FUSION;
D O I
10.1016/j.neunet.2024.106624
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Emotion recognition is an essential but challenging task in human-computer interaction systems due to the distinctive spatial structures and dynamic temporal dependencies associated with each emotion. However, current approaches fail to accurately capture the intricate effects of electroencephalogram (EEG) signals across different brain regions on emotion recognition. Therefore, this paper designs a transformer-based method, denoted by R2G-STLT, which relies on a spatial-temporal transformer encoder with regional to global hierarchical learning that learns the representative spatiotemporal features from the electrode level to the brain-region level. The regional spatial-temporal transformer (RST-Trans) encoder is designed to obtain spatial information and context dependence at the electrode level aiming to learn the regional spatiotemporal features. Then, the global spatial-temporal transformer (GST-Trans) encoder is utilized to extract reliable global spatiotemporal features, reflecting the impact of various brain regions on emotion recognition tasks. Moreover, the multi-head attention mechanism is placed into the GST-Trans encoder to empower it to capture the longrange spatial-temporal information among the brain regions. Finally, subject-independent experiments are conducted on each frequency band of the DEAP, SEED, and SEED-IV datasets to assess the performance of the proposed model. Results indicate that the R2G-STLT model surpasses several state-of-the-art approaches.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] Transformer-Based Multimodal Spatial-Temporal Fusion for Gait Recognition
    Zhang, Jikai
    Ji, Mengyu
    He, Yihao
    Guo, Dongliang
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT XV, 2025, 15045 : 494 - 507
  • [32] RESTHT: relation-enhanced spatial-temporal hierarchical transformer for video captioning
    Zheng, Lihuan
    Xu, Wanru
    Miao, Zhenjiang
    Qiu, Xinxiu
    Gong, Shanshan
    VISUAL COMPUTER, 2025, 41 (01): : 591 - 604
  • [33] Long-term Wind Power Forecasting with Hierarchical Spatial-Temporal Transformer
    Zhang, Yang
    Liu, Lingbo
    Xiong, Xinyu
    Li, Guanbin
    Wang, Guoli
    Lin, Liang
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 6308 - 6316
  • [34] HierGAT: hierarchical spatial-temporal network with graph and transformer for video HOI detection
    Wu, Junxian
    Zhang, Yujia
    Kampffmeyer, Michael
    Pan, Yi
    Zhang, Chenyu
    Sun, Shiying
    Chang, Hui
    Zhao, Xiaoguang
    MULTIMEDIA SYSTEMS, 2025, 31 (01)
  • [35] Emotion recognition using spatial-temporal EEG features through convolutional graph attention network
    Li, Zhongjie
    Zhang, Gaoyan
    Wang, Longbiao
    Wei, Jianguo
    Dang, Jianwu
    JOURNAL OF NEURAL ENGINEERING, 2023, 20 (01)
  • [36] A Neural Network for EEG Emotion Recognition that Combines CNN and Transformer for Multi-scale Spatial-temporal Feature Extraction
    Hu, Zhangfang
    Wu, Haoze
    He, Lingxiao
    IAENG International Journal of Computer Science, 2024, 51 (08) : 1094 - 1104
  • [37] StNet: Local and Global Spatial-Temporal Modeling for Action Recognition
    He, Dongliang
    Zhou, Zhichao
    Gan, Chuang
    Li, Fu
    Liu, Xiao
    Li, Yandong
    Wang, Limin
    Wen, Shilei
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 8401 - 8408
  • [38] Subject-independent emotion recognition of EEG signals using graph attention-based spatial-temporal pattern learning
    Zhu, Yiwen
    Guo, Yeshuang
    Zhu, Wenzhe
    Di, Lare
    Yin, Thong
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 7070 - 7075
  • [39] Learning Complementary Spatial-Temporal Transformer for Video Salient Object Detection
    Liu, Nian
    Nan, Kepan
    Zhao, Wangbo
    Yao, Xiwen
    Han, Junwei
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (08) : 10663 - 10673
  • [40] STMT: A Spatial-Temporal Mesh Transformer for MoCap-Based Action Recognition
    Zhu, Xiaoyu
    Huang, Po-Yao
    Liang, Junwei
    de Melo, Celso M.
    Hauptmann, Alexander
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 1526 - 1536