LAUN Improved StarGAN for Facial Emotion Recognition

被引:14
作者
Wang, Xiaohua [1 ,2 ,3 ]
Gong, Jianqiao [1 ,2 ,3 ]
Hu, Min [1 ,2 ]
Gu, Yu [1 ,2 ]
Ren, Fuji [2 ,4 ]
机构
[1] Hefei Univ Technol, Minist Educ, Key Lab Knowledge Engn Big Data, Hefei 230601, Peoples R China
[2] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Anhui Prov Key Lab Affect Comp & Adv Intelligent, Hefei 230601, Peoples R China
[3] Anhui Prov Key Lab Ind Safety & Emergency Technol, Hefei 230601, Peoples R China
[4] Univ Tokushima, Grad Sch Adv Technol & Sci, Tokushima 7708502, Japan
基金
中国国家自然科学基金;
关键词
Face recognition; Gallium nitride; Generators; Feature extraction; Emotion recognition; Databases; Generative adversarial networks; Facial expression recognition; data enhancement; generative adversarial networks; self-attention;
D O I
10.1109/ACCESS.2020.3021531
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the field of facial expression recognition, deep learning is extensively used. However, insufficient and unbalanced facial training data in available public databases is a major challenge for improving the expression recognition rate. Generative Adversarial Networks (GANs) can produce more one-to-one faces with different expressions, which can be used to enhance databases. StarGAN can perform one-to-many translations for multiple expressions. Compared with original GANs, StarGAN can increase the efficiency of sample generation. Nevertheless, there are some defects in essential areas of the generated face, such as the mouth and the fuzzy side face image generation. To address these limitations, we improved StarGAN to alleviate the defects of images generation by modifying the reconstruction loss and adding the Contextual loss. Meanwhile, we added the Attention U-Net to StarGAN's generator, replacing StarGAN's original generator. Therefore, we proposed the Contextual loss and Attention U-Net (LAUN) improved StarGAN. The U-shape structure and skip connection in Attention U-Net can effectively integrate the details and semantic features of images. The network's attention structure can pay attention to the essential areas of the human face. The experimental results demonstrate that the improved model can alleviate some flaws in the face generated by the original StarGAN. Therefore, it can generate person images with better quality with different poses and expressions. The experiments were conducted on the Karolinska Directed Emotional Faces database, and the accuracy of facial expression recognition is 95.97%, 2.19% higher than that by using StarGAN. Meanwhile, the experiments were carried out on the MMI Facial Expression Database, and the accuracy of expression is 98.30%, 1.21% higher than that by using StarGAN. Moreover, experiment results have better performance based on the LAUN improved StarGAN enhanced databases than those without enhancement.
引用
收藏
页码:161509 / 161518
页数:10
相关论文
共 60 条
[1]  
[Anonymous], representation learning with deep convolutional generative
[2]  
[Anonymous], 2018, PREPRINT
[3]  
[Anonymous], 2017, ARXIV171200433
[4]  
[Anonymous], 2014, ABS14090473 CORR
[5]   ExprADA: Adversarial domain adaptation for facial expression analysis [J].
Bozorgtabar, Behzad ;
Mahapatra, Dwarikanath ;
Thiran, Jean-Philippe .
PATTERN RECOGNITION, 2020, 100
[6]   StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation [J].
Choi, Yunjey ;
Choi, Minje ;
Kim, Munyoung ;
Ha, Jung-Woo ;
Kim, Sunghun ;
Choo, Jaegul .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8789-8797
[7]   Vision and Attention Theory Based Sampling for Continuous Facial Emotion Recognition [J].
Cruz, Albert C. ;
Bhanu, Bir ;
Thakoor, Ninad S. .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2014, 5 (04) :418-431
[8]   Histograms of oriented gradients for human detection [J].
Dalal, N ;
Triggs, B .
2005 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 1, PROCEEDINGS, 2005, :886-893
[9]   A Comprehensive Survey on Pose-Invariant Face Recognition [J].
Ding, Changxing ;
Tao, Dacheng .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2016, 7 (03)
[10]  
Donahue J, 2019, ADV NEUR IN, V32