Emotion recognition based on convolutional neural networks and heterogeneous bio-signal data sources

被引:54
作者
Ngai, Wang Kay [1 ]
Xie, Haoran [2 ]
Zou, Di [3 ]
Chou, Kee-Lee [1 ]
机构
[1] Educ Univ Hong Kong, Dept Asian & Policy Studies, Hong Kong, Peoples R China
[2] Lingnan Univ, Dept Comp & Decis Sci, Hong Kong, Peoples R China
[3] Educ Univ Hong Kong, Dept English Language Educ, Hong Kong, Peoples R China
关键词
Emotion recognition; Electroencephalogram; Arousal-valence model of emotions; 3D convolutional neural network; EEG; MODEL; CLASSIFICATION; REPRESENTATION;
D O I
10.1016/j.inffus.2021.07.007
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Emotion recognition is a crucial application in human-computer interaction. It is usually conducted using facial expressions as the main modality, which might not be reliable. In this study, we proposed a multimodal approach that uses 2-channel electroencephalography (EEG) signals and eye modality in addition to the face modality to enhance the recognition performance. We also studied the use of facial images versus facial depth as the face modality and adapted the common arousal-valence model of emotions and the convolutional neural network, which can model the spatiotemporal information from the modality data for emotion recognition. Extensive experiments were conducted on the modality and emotion data, the results of which showed that our system has high accuracies of 67.8% and 77.0% in valence recognition and arousal recognition, respectively. The proposed method outperformed most state-of-the-art systems that use similar but fewer modalities. Moreover, the use of facial depth has outperformed the use of facial images. The proposed method of emotion recognition has significant potential for integration into various educational applications.
引用
收藏
页码:107 / 117
页数:11
相关论文
共 75 条
[1]  
Antoniou A, 2017, ARXIV171104340
[2]   Neural Networks for Emotion Recognition Based on Eye Tracking Data [J].
Aracena, Claudio ;
Basterrech, Sebastian ;
Snasel, Vaclav ;
Velasquez, Juan .
2015 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC 2015): BIG DATA ANALYTICS FOR HUMAN-CENTRIC SYSTEMS, 2015, :2632-2637
[3]  
Benovoy M, 2008, BIOSIGNALS 2008: PROCEEDINGS OF THE FIRST INTERNATIONAL CONFERENCE ON BIO-INSPIRED SYSTEMS AND SIGNAL PROCESSING, VOL 1, P253
[4]   The pupil as a measure of emotional arousal and autonomic activation [J].
Bradley, Margaret M. ;
Miccoli, Laura ;
Escrig, Miguel A. ;
Lang, Peter J. .
PSYCHOPHYSIOLOGY, 2008, 45 (04) :602-607
[5]  
Carrier P., 2013, FER 2013 FACE DATABA
[6]   Novel Algorithm for Measuring the Complexity of Electroencephalographic Signals in Emotion Recognition [J].
Chen, Dong-Wei ;
Han, Na ;
Chen, Jun-Jie ;
Guo, Hao .
JOURNAL OF MEDICAL IMAGING AND HEALTH INFORMATICS, 2017, 7 (01) :203-210
[7]   Describing the emotional states that are expressed in speech [J].
Cowie, R ;
Cornelius, RR .
SPEECH COMMUNICATION, 2003, 40 (1-2) :5-32
[8]  
De Lemos Jakob., 2008, P MEAS BEH NOLD MAAS, V226, P225
[9]   Collecting Large, Richly Annotated Facial-Expression Databases from Movies [J].
Dhall, Abhinav ;
Goecke, Roland ;
Lucey, Simon ;
Gedeon, Tom .
IEEE MULTIMEDIA, 2012, 19 (03) :34-41
[10]   FaceNet2ExpNet: Regularizing a Deep Face Recognition Net for Expression Recognition [J].
Ding, Hui ;
Zhou, Shaohua Kevin ;
Chellappa, Rama .
2017 12TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2017), 2017, :118-126