Emotion recognition using hierarchical spatial-temporal learning transformer from regional to global brain

被引:4
|
作者
Cheng, Cheng [1 ]
Liu, Wenzhe [2 ]
Feng, Lin [1 ,3 ]
Jia, Ziyu [4 ]
机构
[1] Dalian Univ Technol, Dept Comp Sci & Technol, Dalian, Peoples R China
[2] Huzhou Univ, Sch Informat Engn, Huzhou, Peoples R China
[3] Dalian Minzu Univ, Sch Informat & Commun Engn, Dlian, Peoples R China
[4] Univ Chinese Acad Sci, Chinese Acad Sci, Brainnetome Ctr, Inst Automat, Beijing, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Emotion recognition; Electroencephalogram (EEG); Transformer; Spatiotemporal features; EEG; FUSION;
D O I
10.1016/j.neunet.2024.106624
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Emotion recognition is an essential but challenging task in human-computer interaction systems due to the distinctive spatial structures and dynamic temporal dependencies associated with each emotion. However, current approaches fail to accurately capture the intricate effects of electroencephalogram (EEG) signals across different brain regions on emotion recognition. Therefore, this paper designs a transformer-based method, denoted by R2G-STLT, which relies on a spatial-temporal transformer encoder with regional to global hierarchical learning that learns the representative spatiotemporal features from the electrode level to the brain-region level. The regional spatial-temporal transformer (RST-Trans) encoder is designed to obtain spatial information and context dependence at the electrode level aiming to learn the regional spatiotemporal features. Then, the global spatial-temporal transformer (GST-Trans) encoder is utilized to extract reliable global spatiotemporal features, reflecting the impact of various brain regions on emotion recognition tasks. Moreover, the multi-head attention mechanism is placed into the GST-Trans encoder to empower it to capture the longrange spatial-temporal information among the brain regions. Finally, subject-independent experiments are conducted on each frequency band of the DEAP, SEED, and SEED-IV datasets to assess the performance of the proposed model. Results indicate that the R2G-STLT model surpasses several state-of-the-art approaches.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] From Regional to Global Brain: A Novel Hierarchical Spatial-Temporal Neural Network Model for EEG Emotion Recognition
    Li, Yang
    Zheng, Wenming
    Wang, Lei
    Zong, Yuan
    Cui, Zhen
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2022, 13 (02) : 568 - 578
  • [2] ST-HViT: spatial-temporal hierarchical vision transformer for action recognition
    Xia, Limin
    Fu, Weiye
    PATTERN ANALYSIS AND APPLICATIONS, 2025, 28 (01)
  • [3] TranSkeleton: Hierarchical Spatial-Temporal Transformer for Skeleton-Based Action Recognition
    Liu, Haowei
    Liu, Yongcheng
    Chen, Yuxin
    Yuan, Chunfeng
    Li, Bing
    Hu, Weiming
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (08) : 4137 - 4148
  • [4] Emotion Classification Based on Transformer and CNN for EEG Spatial-Temporal Feature Learning
    Yao, Xiuzhen
    Li, Tianwen
    Ding, Peng
    Wang, Fan
    Zhao, Lei
    Gong, Anmin
    Nan, Wenya
    Fu, Yunfa
    BRAIN SCIENCES, 2024, 14 (03)
  • [5] Focal and Global Spatial-Temporal Transformer for Skeleton-Based Action Recognition
    Gao, Zhimin
    Wang, Peitao
    Lv, Pei
    Jiang, Xiaoheng
    Liu, Qidong
    Wang, Pichao
    Xu, Mingliang
    Li, Wanqing
    COMPUTER VISION - ACCV 2022, PT IV, 2023, 13844 : 155 - 171
  • [6] Local and Global Spatial-Temporal Transformer for skeleton-based action recognition
    Liu, Ruyi
    Chen, Yu
    Gai, Feiyu
    Liu, Yi
    Miao, Qiguang
    Wu, Shuai
    NEUROCOMPUTING, 2025, 634
  • [7] Hierarchical Spatial-Temporal Window Transformer for Pose-Based Rodent Behavior Recognition
    Ru, Zhihao
    Duan, Feng
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 14
  • [8] EEG-Based Emotion Recognition Using Spatial-Temporal Connectivity
    Chu, Wenhao
    Fu, Baole
    Xia, Yuxiao
    Liu, Yinhua
    IEEE ACCESS, 2023, 11 : 92496 - 92504
  • [9] Eye Tracking and Emotion Recognition Using Multiple Spatial-Temporal Networks
    Setianto, Eprian Junan
    Djamal, Esmeralda Contessa
    Nugraha, Fikri
    Kasyidi, Fatan
    2022 INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ITS APPLICATIONS (ICODSA), 2022, : 18 - 23
  • [10] Feature hypergraph representation learning on spatial-temporal correlations for EEG emotion recognition
    Li, Menghang
    Qiu, Min
    Zhu, Li
    Kong, Wanzeng
    COGNITIVE NEURODYNAMICS, 2023, 17 (05) : 1271 - 1281