Emotion Recognition via Multiscale Feature Fusion Network and Attention Mechanism

被引:14
作者
Jiang, Yiye [1 ]
Xie, Songyun [1 ]
Xie, Xinzhou [1 ]
Cui, Yujie [1 ]
Tang, Hao [1 ]
机构
[1] Northwestern Polytech Univ, Sch Elect & Informat, Xian 710072, Shaanxi, Peoples R China
关键词
Feature extraction; Electroencephalography; Emotion recognition; Convolution; Brain modeling; Deep learning; Task analysis; Attention mechanism; deep learning; electroencephalogram (EEG); emotion recognition; feature fusion; spatial-temporal feature; CONVOLUTIONAL NEURAL-NETWORK; FEATURE-EXTRACTION; EEG; CLASSIFICATION;
D O I
10.1109/JSEN.2023.3265688
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Traditional manual feature-based machine learning methods and deep learning networks have been used for electroencephalogram (EEG)-based emotion recognition in recent years. However, some existing studies ignore the low signal-to-noise ratio and the fact that each subject has unique EEG traits, which suffer from low recognition accuracy and poor robustness. To solve these problems, we propose a novel attention mechanism-based multiscale feature fusion network (AM-MSFFN) that considers high-level features at different scales to improve the generalization of the model for different subjects. Specifically, we first utilize a spatial-temporal convolutional block to extract temporal and spatial features of EEG signals sequentially. Subsequently, considering the sampling rate of EEG signals, the multiscale separable convolutions are designed for capturing emotional state-related information, to better combine and output feature mapping relationships. Convolutional module attention mechanism (CBAM) is applied after point-wise convolution, to better handle EEG variations of different subjects and the key information which facilitates classification. In addition, we adopt a preprocessing module based on data augmentation and data alignment to improve the quality of the training samples. Moreover, ablation studies show that the proposed attention mechanism and multiscale separable convolution contribute significant and consistent gain to the performance of our AM-MSFFN model. To verify the effectiveness of the proposed algorithm, we conducted extensive experiments on the DEAP dataset and SEED. The average accuracies achieve 99.479% and 99.297% for arousal and valence, respectively. The results demonstrated the feasibility of the proposed method.
引用
收藏
页码:10790 / 10800
页数:11
相关论文
共 55 条
[51]   EEG emotion recognition using fusion model of graph convolutional neural networks and LSTM [J].
Yin, Yongqiang ;
Zheng, Xiangwei ;
Hu, Bin ;
Zhang, Yuang ;
Cui, Xinchun .
APPLIED SOFT COMPUTING, 2021, 100
[52]   Emotion recognition using multi-modal data and machine learning techniques: A tutorial and review [J].
Zhang, Jianhua ;
Yin, Zhong ;
Chen, Peng ;
Nichele, Stefano .
INFORMATION FUSION, 2020, 59 :103-126
[53]   GCB-Net: Graph Convolutional Broad Network and Its Application in Emotion Recognition [J].
Zhang, Tong ;
Wang, Xuehan ;
Xu, Xiangmin ;
Chen, C. L. Philip .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2022, 13 (01) :379-388
[54]   Expression EEG Multimodal Emotion Recognition Method Based on the Bidirectional LSTM and Attention Mechanism [J].
Zhao, Yifeng ;
Chen, Deyun .
COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE, 2021, 2021
[55]   Investigating Critical Frequency Bands and Channels for EEG-Based Emotion Recognition with Deep Neural Networks [J].
Zheng, Wei-Long ;
Lu, Bao-Liang .
IEEE TRANSACTIONS ON AUTONOMOUS MENTAL DEVELOPMENT, 2015, 7 (03) :162-175