Continuous Emotion Recognition with Spatiotemporal Convolutional Neural Networks

被引:5
作者
Teixeira, Thomas [1 ]
Granger, Eric [1 ]
Lameiras Koerich, Alessandro [1 ]
机构
[1] Univ Quebec, Ecole Technol Super, 1100 Rue Notre Dame Ouest, Montreal, PQ H3C 1K3, Canada
来源
APPLIED SCIENCES-BASEL | 2021年 / 11卷 / 24期
基金
加拿大自然科学与工程研究理事会;
关键词
facial expression recognition; deep learning; convolutional recurrent neural networks; inflated 3D CNNs; dimensional emotion representation; long short-term memory; FACIAL EXPRESSIONS; DEEP; FEATURES; IMAGE; FACE;
D O I
10.3390/app112411738
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Facial expressions are one of the most powerful ways to depict specific patterns in human behavior and describe the human emotional state. However, despite the impressive advances of affective computing over the last decade, automatic video-based systems for facial expression recognition still cannot correctly handle variations in facial expression among individuals as well as cross-cultural and demographic aspects. Nevertheless, recognizing facial expressions is a difficult task, even for humans. This paper investigates the suitability of state-of-the-art deep learning architectures based on convolutional neural networks (CNNs) to deal with long video sequences captured in the wild for continuous emotion recognition. For such an aim, several 2D CNN models that were designed to model spatial information are extended to allow spatiotemporal representation learning from videos, considering a complex and multi-dimensional emotion space, where continuous values of valence and arousal must be predicted. We have developed and evaluated convolutional recurrent neural networks, combining 2D CNNs and long short term-memory units and inflated 3D CNN models, which are built by inflating the weights of a pre-trained 2D CNN model during fine-tuning, using application-specific videos. Experimental results on the challenging SEWA-DB dataset have shown that these architectures can effectively be fine-tuned to encode spatiotemporal information from successive raw pixel images and achieve state-of-the-art results on such a dataset.
引用
收藏
页数:21
相关论文
共 75 条
[21]   Audio and Face Video Emotion Recognition in the Wild using Deep Neural Networks and Small Datasets [J].
Ding, Wan ;
Xu, Mingyu ;
Huang, Dongyan ;
Lin, Weisi ;
Dong, Minghui ;
Yu, Xinguo ;
Li, Haizhou .
ICMI'16: PROCEEDINGS OF THE 18TH ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2016, :506-513
[22]   Learning Spatiotemporal Features with 3D Convolutional Networks [J].
Du Tran ;
Bourdev, Lubomir ;
Fergus, Rob ;
Torresani, Lorenzo ;
Paluri, Manohar .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :4489-4497
[23]   STRONG EVIDENCE FOR UNIVERSALS IN FACIAL EXPRESSIONS - A REPLY TO RUSSELLS MISTAKEN CRITIQUE [J].
EKMAN, P .
PSYCHOLOGICAL BULLETIN, 1994, 115 (02) :268-287
[24]   CONSTANTS ACROSS CULTURES IN FACE AND EMOTION [J].
EKMAN, P ;
FRIESEN, WV .
JOURNAL OF PERSONALITY AND SOCIAL PSYCHOLOGY, 1971, 17 (02) :124-&
[25]   Video-Based Emotion Recognition using CNN-RNN and C3D Hybrid Networks [J].
Fan, Yin ;
Lu, Xiangju ;
Li, Dian ;
Liu, Yuanliu .
ICMI'16: PROCEEDINGS OF THE 18TH ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2016, :445-450
[26]   Time Perception and Dynamics of Facial Expressions of Emotions [J].
Fayolle, Sophie L. ;
Droit-Volet, Sylvie .
PLOS ONE, 2014, 9 (05)
[27]   Debate about universal facial expressions goes big [J].
Feldman Barrett, Lisa .
NATURE, 2021, 589 (7841) :202-203
[28]   Local Learning With Deep and Handcrafted Features for Facial Expression Recognition [J].
Georgescu, Mariana-Iuliana ;
Ionescu, Radu Tudor ;
Popescu, Marius .
IEEE ACCESS, 2019, 7 :64827-64836
[29]  
Goodfellow Ian J., 2013, Neural Information Processing. 20th International Conference, ICONIP 2013. Proceedings: LNCS 8228, P117, DOI 10.1007/978-3-642-42051-1_16
[30]   Deep Neural Networks with Relativity Learning for Facial Expression Recognition [J].
Guo, Yanan ;
Tao, Dapeng ;
Yu, Jun ;
Xiong, Hao ;
Li, Yaotang ;
Tao, Dacheng .
2016 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2016,