Emotion Recognition using Multimodal Residual LSTM Network

被引:149
作者
Ma, Jiaxin [1 ]
Tang, Hao [2 ]
Zheng, Wei-Long [3 ]
Lu, Bao-Liang [2 ]
机构
[1] OMRON SINIC X Corp, Tokyo, Japan
[2] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
[3] Harvard Med Sch, Massachusetts Gen Hosp, Boston, MA 02115 USA
来源
PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19) | 2019年
基金
中国国家自然科学基金;
关键词
Multimodal emotion recognition; long-short-term memory network; electroencephalography; EEG; SIGNALS;
D O I
10.1145/3343031.3350871
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Various studies have shown that the temporal information captured by conventional long-short-term memory (LSTM) networks is very useful for enhancing multimodal emotion recognition using encephalography (EEG) and other physiological signals. However, the dependency among multiple modalities and high-level temporal-feature learning using deeper LSTM networks is yet to be investigated. Thus, we propose a multimodal residual LSTM (MMResLSTM) network for emotion recognition. The MMResLSTM network shares the weights across the modalities in each LSTM layer to learn the correlation between the EEG and other physiological signals. It contains both the spatial shortcut paths provided by the residual network and temporal shortcut paths provided by LSTM for efficiently learning emotion-related high-level features. The proposed network was evaluated using a publicly available dataset for EEG-based emotion recognition, DEAP. The experimental results indicate that the proposed MMResLSTM network yielded a promising result, with a classification accuracy of 92.87% for arousal and 92.30% for valence.
引用
收藏
页码:176 / 183
页数:8
相关论文
共 40 条
[1]  
[Anonymous], 2017, INT J ADV COMPUTER S
[2]  
[Anonymous], 2015, ARXIV PREPRINT ARXIV
[3]  
[Anonymous], 2016, ICLR C
[4]  
[Anonymous], NEURAL COMPUT
[5]  
Ba L. J., 2016, Layer Nor- malization
[6]   Emotion Recognition Based on High-Resolution EEG Recordings and Reconstructed Brain Sources [J].
Becker, Hanna ;
Fleureau, Julien ;
Guillotel, Philippe ;
Wendling, Fabrice ;
Merlet, Isabelle ;
Albera, Laurent .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2020, 11 (02) :244-257
[7]  
Busso C., 2004, Proceedings of the 6th international conference on Multimodal interfaces, P205, DOI [10.1145/1027933.1027968, DOI 10.1145/1027933.1027968]
[8]  
Candra H, 2015, IEEE ENG MED BIO, P7250, DOI 10.1109/EMBC.2015.7320065
[9]   RelationLines: Visual Reasoning of Egocentric Relations from Heterogeneous Urban Data [J].
Chen, Wei ;
Xia, Jing ;
Wang, Xumeng ;
Wang, Yi ;
Chen, Jun ;
Chang, Liang .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2019, 10 (01)
[10]   Optimized Echo State Network with Intrinsic Plasticity for EEG-Based Emotion Recognition [J].
Fourati, Rahma ;
Ammar, Boudour ;
Aouiti, Chaouki ;
Sanchez-Medina, Javier ;
Alimi, Adel M. .
NEURAL INFORMATION PROCESSING (ICONIP 2017), PT II, 2017, 10635 :718-727