HEU Emotion: a large-scale database for multimodal emotion recognition in the wild

被引:0
作者
Jing Chen
Chenhui Wang
Kejun Wang
Chaoqun Yin
Cong Zhao
Tao Xu
Xinyi Zhang
Ziqiang Huang
Meichen Liu
Tao Yang
机构
[1] Harbin Engineering University,College of Intelligent Systems Science and Engineering
[2] UCLA Department of Statistics,undefined
来源
Neural Computing and Applications | 2021年 / 33卷
关键词
Emotion recognition in the wild; Multimodal; Facial expression; Body posture; Emotional speech;
D O I
暂无
中图分类号
学科分类号
摘要
The study of affective computing in the wild setting is underpinned by databases. Existing multimodal emotion databases in the real-world conditions are few and small, with a limited number of subjects and expressed in a single language. To meet this requirement, we collected, annotated, and prepared to release a new natural state video database (called HEU Emotion). HEU Emotion contains a total of 19,004 video clips, which is divided into two parts according to the data source. The first part contains videos downloaded from Tumblr, Google, and Giphy, including 10 emotions and two modalities (facial expression and body posture). The second part includes corpus taken manually from movies, TV series, and variety shows, consisting of 10 emotions and three modalities (facial expression, body posture, and emotional speech). HEU Emotion is by far the most extensive multimodal emotional database with 9951 subjects. In order to provide a benchmark for emotion recognition, we used many conventional machine learning and deep learning methods to evaluate HEU Emotion. We proposed a multimodal attention module to fuse multimodal features adaptively. After multimodal fusion, the recognition accuracies for the two parts increased by 2.19% and 4.01%, respectively, over those of single-modal facial expression recognition.
引用
收藏
页码:8669 / 8685
页数:16
相关论文
共 50 条
[1]  
Busso C(2008)Iemocap: Interactive emotional dyadic motion capture database Language resources and evaluation 42 335-339
[2]  
Bulut M(2012)Collecting large, richly annotated facial-expression databases from movies IEEE MultiMed 19 34-41
[3]  
Lee CC(1993)Facial expression and emotion Am psychol 48 384-140
[4]  
Kazemzadeh A(2003)Emotions revealed: Recognizing faces and feelings to improve communication and emotional life Holt Paperback 128 140-64836
[5]  
Mower E(2019)Local learning with deep and handcrafted features for facial expression recognition IEEE Access 7 64827-924
[6]  
Kim S(2017)Cheavd: a chinese natural emotional audio-visual database J Amb Intell Human Comput 8 913-476
[7]  
Chang JN(2002)Gabor feature based classification using the enhanced fisher linear discriminant model for face recognition IEEE Trans Image process 11 467-191
[8]  
Lee S(2018)The ryerson audio-visual database of emotional speech and song (ravdess): A dynamic, multimodal set of facial and vocal expressions in north american english PloS one 13 e0196391-17
[9]  
Narayanan SS(2011)Evidence for training the ability to read microexpressions of emotion Motivation Emotion 35 181-31
[10]  
Dhall A(1976)Hearing lips and seeing voices Nature 264 746-327