Emotion Recognition using Multimodal Residual LSTM Network

被引:149
作者
Ma, Jiaxin [1 ]
Tang, Hao [2 ]
Zheng, Wei-Long [3 ]
Lu, Bao-Liang [2 ]
机构
[1] OMRON SINIC X Corp, Tokyo, Japan
[2] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
[3] Harvard Med Sch, Massachusetts Gen Hosp, Boston, MA 02115 USA
来源
PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19) | 2019年
基金
中国国家自然科学基金;
关键词
Multimodal emotion recognition; long-short-term memory network; electroencephalography; EEG; SIGNALS;
D O I
10.1145/3343031.3350871
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Various studies have shown that the temporal information captured by conventional long-short-term memory (LSTM) networks is very useful for enhancing multimodal emotion recognition using encephalography (EEG) and other physiological signals. However, the dependency among multiple modalities and high-level temporal-feature learning using deeper LSTM networks is yet to be investigated. Thus, we propose a multimodal residual LSTM (MMResLSTM) network for emotion recognition. The MMResLSTM network shares the weights across the modalities in each LSTM layer to learn the correlation between the EEG and other physiological signals. It contains both the spatial shortcut paths provided by the residual network and temporal shortcut paths provided by LSTM for efficiently learning emotion-related high-level features. The proposed network was evaluated using a publicly available dataset for EEG-based emotion recognition, DEAP. The experimental results indicate that the proposed MMResLSTM network yielded a promising result, with a classification accuracy of 92.87% for arousal and 92.30% for valence.
引用
收藏
页码:176 / 183
页数:8
相关论文
共 40 条
[21]  
Madigan JE, 2016, LEAN FOR THE CASH-STRAPPED LEADER: THE PATH TO GROWTH AND PROFITABILITY, P163
[22]   Emotional State Recognition with EEG Signals Using Subject Independent Approach [J].
Pandey, Pallavi ;
Seeja, K. R. .
DATA SCIENCE AND BIG DATA ANALYTICS, 2019, 16 :117-124
[23]   A Mutual Information Based Adaptive Windowing of Informative EEG for Emotion Recognition [J].
Piho, Laura ;
Tjahjadi, Tardi .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2020, 11 (04) :722-735
[24]   Multi-view Emotion Recognition Using Deep Canonical Correlation Analysis [J].
Qiu, Jie-Lin ;
Liu, Wei ;
Lu, Bao-Liang .
NEURAL INFORMATION PROCESSING (ICONIP 2018), PT V, 2018, 11305 :221-231
[25]  
Ren J, 2016, AAAI CONF ARTIF INTE, P3581
[26]   Integrating cross-scale analysis in the spatial and temporal domains for classification of behavioral movement [J].
Soleymani, Ali ;
Cachat, Jonathan ;
Robinson, Kyle ;
Dodge, Somayeh ;
Kalueff, Allan V. ;
Weibel, Robert .
JOURNAL OF SPATIAL INFORMATION SCIENCE, 2014, (08) :1-25
[27]   A Multimodal Database for Affect Recognition and Implicit Tagging [J].
Soleymani, Mohammad ;
Lichtenauer, Jeroen ;
Pun, Thierry ;
Pantic, Maja .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2012, 3 (01) :42-55
[28]  
Srivastava RK., 2015, P 28 INT C NEURAL IN, P2377, DOI DOI 10.48550/ARXIV.1507.06228
[29]   DeepSleepNet: A Model for Automatic Sleep Stage Scoring Based on Raw Single-Channel EEG [J].
Supratak, Akara ;
Dong, Hao ;
Wu, Chao ;
Guo, Yike .
IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2017, 25 (11) :1998-2008
[30]  
Szegedy C, 2015, PROC CVPR IEEE, P1, DOI 10.1109/CVPR.2015.7298594