Emotion Classification Based on Transformer and CNN for EEG Spatial-Temporal Feature Learning

被引:9
|
作者
Yao, Xiuzhen [1 ,2 ]
Li, Tianwen [2 ,3 ]
Ding, Peng [1 ,2 ]
Wang, Fan [1 ,2 ]
Zhao, Lei [2 ,3 ]
Gong, Anmin [4 ]
Nan, Wenya [5 ]
Fu, Yunfa [1 ,2 ]
机构
[1] Kunming Univ Sci & Technol, Fac Informat Engn & Automat, Kunming 650500, Peoples R China
[2] Kunming Univ Sci & Technol, Brain Cognit & Brain Comp Intelligence Integrat Gr, Kunming 650500, Peoples R China
[3] Kunming Univ Sci & Technol, Fac Sci, Kunming 650500, Peoples R China
[4] Chinese Peoples Armed Police Force Engn Univ, Sch Informat Engn, Xian 710086, Peoples R China
[5] Shanghai Normal Univ, Coll Educ, Dept Psychol, Shanghai 200234, Peoples R China
基金
中国国家自然科学基金;
关键词
EEG; emotion classification; transformer; CNN; multi-head attention;
D O I
10.3390/brainsci14030268
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Objectives: The temporal and spatial information of electroencephalogram (EEG) signals is crucial for recognizing features in emotion classification models, but it excessively relies on manual feature extraction. The transformer model has the capability of performing automatic feature extraction; however, its potential has not been fully explored in the classification of emotion-related EEG signals. To address these challenges, the present study proposes a novel model based on transformer and convolutional neural networks (TCNN) for EEG spatial-temporal (EEG ST) feature learning to automatic emotion classification. Methods: The proposed EEG ST-TCNN model utilizes position encoding (PE) and multi-head attention to perceive channel positions and timing information in EEG signals. Two parallel transformer encoders in the model are used to extract spatial and temporal features from emotion-related EEG signals, and a CNN is used to aggregate the EEG's spatial and temporal features, which are subsequently classified using Softmax. Results: The proposed EEG ST-TCNN model achieved an accuracy of 96.67% on the SEED dataset and accuracies of 95.73%, 96.95%, and 96.34% for the arousal-valence, arousal, and valence dimensions, respectively, for the DEAP dataset. Conclusions: The results demonstrate the effectiveness of the proposed ST-TCNN model, with superior performance in emotion classification compared to recent relevant studies. Significance: The proposed EEG ST-TCNN model has the potential to be used for EEG-based automatic emotion recognition.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] EEG Classification with Transformer-Based Models
    Sun, Jiayao
    Xie, Jin
    Zhou, Huihui
    2021 IEEE 3RD GLOBAL CONFERENCE ON LIFE SCIENCES AND TECHNOLOGIES (IEEE LIFETECH 2021), 2021, : 92 - 93
  • [22] IEEG-CT: A CNN and Transformer Based Method for Intracranial EEG Signal Classification
    Yu, Mengxin
    Zhang, Yuang
    Liu, Haihui
    Wu, Xiaona
    Du, Mingsen
    Liu, Xiaojie
    NEURAL INFORMATION PROCESSING, ICONIP 2023, PT III, 2024, 14449 : 549 - 563
  • [23] A Spatial-Temporal Transformer Architecture Using Multi-Channel Signals for Sleep Stage Classification
    Yao, Haotian
    Liu, Tao
    Zou, Ruiyang
    Ding, Shengnan
    Xu, Yan
    IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2023, 31 : 3353 - 3362
  • [24] A Multi-Modal Emotion Recognition System Based on CNN-Transformer Deep Learning Technique
    Karatay, Busra
    Bestepe, Deniz
    Sailunaz, Kashfia
    Ozyer, Tansel
    Alhajj, Reda
    2022 7TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND MACHINE LEARNING APPLICATIONS (CDMA 2022), 2022, : 145 - 150
  • [25] Learning Complementary Spatial-Temporal Transformer for Video Salient Object Detection
    Liu, Nian
    Nan, Kepan
    Zhao, Wangbo
    Yao, Xiwen
    Han, Junwei
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (08) : 10663 - 10673
  • [26] Parallel Spatial-Temporal Self-Attention CNN-Based Motor Imagery Classification for BCI
    Liu, Xiuling
    Shen, Yonglong
    Liu, Jing
    Yang, Jianli
    Xiong, Peng
    Lin, Feng
    FRONTIERS IN NEUROSCIENCE, 2020, 14
  • [27] TRANSTL: SPATIAL-TEMPORAL LOCALIZATION TRANSFORMER FOR MULTI-LABEL VIDEO CLASSIFICATION
    Wu, Hongjun
    Li, Mengzhu
    Liu, Yongcheng
    Liu, Hongzhe
    Xu, Cheng
    Li, Xuewei
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 1965 - 1969
  • [28] ASTDF-Net: Attention-Based Spatial-Temporal Dual-Stream Fusion Network for EEG-Based Emotion Recognition
    Gong, Peiliang
    Jia, Ziyu
    Wang, Pengpai
    Zhou, Yueying
    Zhang, Daoqiang
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 883 - 892
  • [29] Learning spatial-temporal features for video copy detection by the combination of CNN and RNN
    Hu, Yaocong
    Lu, Xiaobo
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2018, 55 : 21 - 29
  • [30] An Extended Common Spatial Pattern Framework for EEG-Based Emotion Classification
    Chen, Jingxia
    Jiang, Dongmei
    Zhang, Yanning
    ADVANCES IN BRAIN INSPIRED COGNITIVE SYSTEMS, BICS 2018, 2018, 10989 : 282 - 292