Multimodal emotion recognition based on manifold learning and convolution neural network

被引:0
作者
Yong Zhang
Cheng Cheng
YiDie Zhang
机构
[1] Huzhou University,School of Information Engineering
[2] Liaoning Normal University,School of Computer and Information Technology
来源
Multimedia Tools and Applications | 2022年 / 81卷
关键词
Multimodal; Emotion recognition; Physiological signals; Manifold learning; Convolution neural network;
D O I
暂无
中图分类号
学科分类号
摘要
Multimodal emotion recognition task based on physiological signals is becoming a research hotspot. Traditional methods need to design and extract a series of features from single-channel or multi-channel physiological signals on the basis of extensive domain knowledge. These methods cannot make full use of the relevant information among channels, and the emotion recognition of a single modality cannot fully express the emotional state. This paper proposes a multimodal emotion recognition model based on manifold learning and a convolutional neural network (CNN). The electroencephalograph (EEG) signals are combined with peripheral physiological signals and eye movement signals respectively, the multivariate synchrosqueezing transform (MSST) is used to simulate the joint oscillation structure of multi-channel signals, and then the related feature parameters are extracted and fused into feature vectors. The proposed method finds the corresponding low dimensional embedding features for given high-dimensional features by an improved manifold learning method, which feeds into the deep convolutional neural network (DCNN) model for emotion recognition. We perform extensive four-category experiments on the dimensions of arousal and valence. Results indicate that our proposed model achieves average accuracies of 90.05% and 88.17% on the DEAP and MAHNOB-HCI datasets respectively, which both receive better performances than most of the compared studies, verifying the effectiveness of the model.
引用
收藏
页码:33253 / 33268
页数:15
相关论文
共 129 条
[1]  
Ahrabian A(2015)Synchrosqueezing-based time-frequency analysis of multivariate data Signal Process 106 331-341
[2]  
Looney D(2016)Improving BCI-based emotion recognition by combining EEG feature selection and kernel classifiers Expert Syst Appl 47 35-41
[3]  
Stanković L(2020)Emotion recognition from multi-channel EEG signals by exploiting the deep belief-conditional random field framework IEEE Access 8 33002-33012
[4]  
Mandic DP(2019)Emotion recognition from multiband EEG signals using CapsNet Sensor 19 2212-32934
[5]  
Atkinson J(2020)Speech emotion recognition using convolutional neural network and long-short term memory Multimed Tools Appl 79 32917-954
[6]  
Campos D(2021)A channel-fused dense convolutional network for EEG-based emotion recognition IEEE Trans Cogn Develop Syst 13 945-2274
[7]  
Chao H(2019)Cross-subject emotion recognition using flexible analytic wavelet transform from EEG signals IEEE Sensors J 19 2266-8240
[8]  
Liu YL(2020)Development of a real-time emotion recognition system using facial expressions and EEG based on machine learning and deep neural network methods Informatics in Medicine Unlocked 20 100372-3271
[9]  
Chao H(2021)Video multimodal emotion recognition based on Bi-GRU and attention fusion Multimed Tools Appl 80 8213-440
[10]  
Dong L(2020)Multimodal emotion recognition based on ensemble convolutional neural network IEEE Access 8 3265-31