The Fusion of Electroencephalography and Facial Expression for Continuous Emotion Recognition

被引:49
作者
Li, Dahua [1 ]
Wang, Zhe [1 ]
Wang, Chuhan [1 ]
Liu, Shuang [1 ]
Chi, Wenhao [1 ]
Dong, Enzeng [1 ]
Song, Xiaolin [2 ]
Gao, Qiang [1 ]
Song, Yu [1 ]
机构
[1] Tianjin Univ Technol, Tianjin Key Lab Control Theory & Applicat Complic, Tianjin 300384, Peoples R China
[2] Tianjin Univ Technol, Engn Training Ctr, Tianjin 300384, Peoples R China
关键词
Emotion recognition; Electroencephalography; Feature extraction; Brain modeling; Motion pictures; Human computer interaction; Electrodes; Continuous emotion recognition; EEG; facial expressions; signal processing; decision level fusion; temporal dynamics; NEURAL-NETWORK; EEG;
D O I
10.1109/ACCESS.2019.2949707
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, the study of emotion recognition has received increasing attentions by the rapid development of noninvasive sensor technologies, machine learning algorithms and compute capability of computers. Compared with single modal emotion recognition, the multimodal paradigm introduces complementary information for emotion recognition. Hence, in this work, we presented a decision level fusion framework for detecting emotions continuously by fusing the Electroencephalography (EEG) and facial expressions. Three types of movie clips (positive, negative, and neutral) were utilized to elicit specific emotions of subjects, the EEG and facial expression signals were recorded simultaneously. The power spectrum density (PSD) features of EEG were extracted by time-frequency analysis, and then EEG features were selected for regression. For the facial expression, the facial geometric features were calculated by facial landmark localization. Long short-term memory networks (LSTM) were utilized to accomplish the decision level fusion and captured temporal dynamics of emotions. The results have shown that the proposed method achieved outstanding performance for continuous emotion recognition, and it yields 0.625 0.029 of concordance correlation coefficient (CCC). From the results, the fusion of two modalities outperformed EEG and facial expression separately. Furthermore, different numbers of time-steps of LSTM was applied to analyze the temporal dynamic capturing.
引用
收藏
页码:155724 / 155736
页数:13
相关论文
共 50 条
[1]  
[Anonymous], 2016, P 25 INT JOINT C ART
[2]  
Banda N, 2015, INT CONF AFFECT, P380, DOI 10.1109/ACII.2015.7344599
[3]  
BRADLEY J, 2007, HDB EMOTION ELICITAT, P19
[4]   MEASURING EMOTION - THE SELF-ASSESSMENT MANNEQUIN AND THE SEMANTIC DIFFERENTIAL [J].
BRADLEY, MM ;
LANG, PJ .
JOURNAL OF BEHAVIOR THERAPY AND EXPERIMENTAL PSYCHIATRY, 1994, 25 (01) :49-59
[5]   IEMOCAP: interactive emotional dyadic motion capture database [J].
Busso, Carlos ;
Bulut, Murtaza ;
Lee, Chi-Chun ;
Kazemzadeh, Abe ;
Mower, Emily ;
Kim, Samuel ;
Chang, Jeannette N. ;
Lee, Sungbok ;
Narayanan, Shrikanth S. .
LANGUAGE RESOURCES AND EVALUATION, 2008, 42 (04) :335-359
[6]  
Chaparro V, 2018, IEEE ENG MED BIO, P530, DOI 10.1109/EMBC.2018.8512407
[7]  
Chollet F., 2015, Keras: Simple Flexible Powerful
[8]  
Cid Felipe., 2015, Proceedings of FinE, P1
[9]   A Review and Meta-Analysis of Multimodal Affect Detection Systems [J].
D'Mello, Sidney K. ;
Kory, Jacqueline .
ACM COMPUTING SURVEYS, 2015, 47 (03)
[10]   Affective neuroscience and psychophysiology: Toward a synthesis [J].
Davidson, RJ .
PSYCHOPHYSIOLOGY, 2003, 40 (05) :655-665