A multi-stage dynamical fusion network for multimodal emotion recognition

被引:20
作者
Chen, Sihan [2 ]
Tang, Jiajia [1 ]
Zhu, Li [1 ]
Kong, Wanzeng [1 ]
机构
[1] Hangzhou Dianzi Univ, Coll Comp Sci, Hangzhou, Peoples R China
[2] Hangzhou Dianzi Univ, HDU ITMO Joint Inst, Hangzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Physiological signals; Emotion recognition; Multimodal dynamic fusion; Multi-stage fusion;
D O I
10.1007/s11571-022-09851-w
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
In recent years, emotion recognition using physiological signals has become a popular research topic. Physiological signal can reflect the real emotional state for individual which is widely applied to emotion recognition. Multimodal signals provide more discriminative information compared with single modal which arose the interest of related researchers. However, current studies on multimodal emotion recognition normally adopt one-stage fusion method which results in the overlook of cross-modal interaction. To solve this problem, we proposed a multi-stage multimodal dynamical fusion network (MSMDFN). Through the MSMDFN, the joint representation based on cross-modal correlation is obtained. Initially, the latent and essential interactions among various features extracted independently from multiple modalities are explored based on specific manner. Subsequently, the multi-stage fusion network is designed to split the fusion procedure into multi-stages using the correlation observed before. This allows us to exploit much more fine-grained unimodal, bimodal and trimodal intercorrelations. For evaluation, the MSMDFN was verified on multimodal benchmark DEAP. The experiments indicate that our method outperforms the related one-stage multi-modal emotion recognition works.
引用
收藏
页码:671 / 680
页数:10
相关论文
共 33 条
[1]   Detecting Naturalistic Expressions of Nonbasic Affect Using Physiological Signals [J].
AlZoubi, Omar ;
D'Mello, Sidney K. ;
Calvo, Rafael A. .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2012, 3 (03) :298-310
[2]   Bipolar oscillations between positive and negative mood states in a computational model of Basal Ganglia [J].
Balasubramani, Pragathi Priyadharsini ;
Chakravarthy, V. Srinivasa .
COGNITIVE NEURODYNAMICS, 2020, 14 (02) :181-202
[3]  
Chen J, 2016, IEEE INT C BIOINFORM, P470, DOI 10.1109/BIBM.2016.7822562
[4]  
Chen J, 2015, IEEE INT C BIOINFORM, P395, DOI 10.1109/BIBM.2015.7359713
[5]   Electroencephalogram-based emotion assessment system using ontology and data mining techniques [J].
Chen, Jing ;
Hu, Bin ;
Moore, Philip ;
Zhang, Xiaowei ;
Ma, Xu .
APPLIED SOFT COMPUTING, 2015, 30 :663-674
[6]   Xception: Deep Learning with Depthwise Separable Convolutions [J].
Chollet, Francois .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1800-1807
[7]   ASYMMETRICAL BRAIN ACTIVITY DISCRIMINATES BETWEEN POSITIVE AND NEGATIVE AFFECTIVE STIMULI IN HUMAN INFANTS [J].
DAVIDSON, RJ ;
FOX, NA .
SCIENCE, 1982, 218 (4578) :1235-1237
[8]   Investigating time-varying functional connectivity derived from the Jackknife Correlation method for distinguishing between emotions in fMRI data [J].
Ghahari, Shabnam ;
Farahani, Naemeh ;
Fatemizadeh, Emad ;
Nasrabadi, Ali Motie .
COGNITIVE NEURODYNAMICS, 2020, 14 (04) :457-471
[9]   EEG spectral powers and source localization in depressing, sad, and fun music videos focusing on gender differences [J].
Goshvarpour, Atefeh ;
Goshvarpour, Ateke .
COGNITIVE NEURODYNAMICS, 2019, 13 (02) :161-173
[10]   Deep Neural Networks for Acoustic Modeling in Speech Recognition [J].
Hinton, Geoffrey ;
Deng, Li ;
Yu, Dong ;
Dahl, George E. ;
Mohamed, Abdel-rahman ;
Jaitly, Navdeep ;
Senior, Andrew ;
Vanhoucke, Vincent ;
Patrick Nguyen ;
Sainath, Tara N. ;
Kingsbury, Brian .
IEEE SIGNAL PROCESSING MAGAZINE, 2012, 29 (06) :82-97