Efficient Facial Expression Recognition Algorithm Based on Hierarchical Deep Neural Network Structure

被引:170
作者
Kim, Ji-Hae [1 ]
Kim, Byung-Gyu [1 ]
Roy, Partha Pratim [2 ]
Jeong, Da-Mi [1 ]
机构
[1] Sookmyung Womens Univ, Dept IT Engn, Seoul, South Korea
[2] IIT Roorkee, Sch Comp Sci & Engn, Roorkee, Uttar Pradesh, India
基金
新加坡国家研究基金会;
关键词
Artificial intelligence (AI); facial expression recognition (FER); emotion recognition; deep learning; LBP feature; geometric feature; convolutional neural network (CNN); FEATURES;
D O I
10.1109/ACCESS.2019.2907327
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the continued development of artificial intelligence (AI) technology, research on interaction technology has become more popular. Facial expression recognition (FER) is an important type of visual information that can be used to understand a human's emotional situation. In particular, the importance of AI systems has recently increased due to advancements in research on AI systems applied to AI robots. In this paper, we propose a new scheme for FER system based on hierarchical deep learning. The feature extracted from the appearance feature-based network is fused with the geometric feature in a hierarchical structure. The appearance feature-based network extracts holistic features of the face using the preprocessed LBP image, whereas the geometric feature-based network learns the coordinate change of action units (AUs) landmark, which is a muscle that moves mainly when making facial expressions. The proposed method combines the result of the softmax function of two features by considering the error associated with the second highest emotion (Top-2) prediction result. In addition, we propose a technique to generate facial images with neutral emotion using the autoencoder technique. By this technique, we can extract the dynamic facial features between the neutral and emotional images without sequence data. We compare the proposed algorithm with the other recent algorithms for CK+ and JAFFE dataset, which are typically considered to be verified datasets in the facial expression recognition. The ten-fold cross validation results show 96.46% of accuracy in the CK+ dataset and 91.27% of accuracy in the JAFFE dataset. When comparing with other methods, the result of the proposed hierarchical deep network structure shows up to about 3% of the accuracy improvement and 1.3% of average improvement in CK+ dataset, respectively. In JAFFE datasets, up to about 7% of the accuracy is enhanced, and the average improvement is verified by about 1.5%.
引用
收藏
页码:41273 / 41285
页数:13
相关论文
共 36 条
[1]   Survey on RGB, 3D, Thermal, and Multimodal Approaches for Facial Expression Recognition: History, Trends, and Affect-Related Applications [J].
Adrian Corneanu, Ciprian ;
Oliu Simon, Marc ;
Cohn, Jeffrey F. ;
Escalera Guerrero, Sergio .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2016, 38 (08) :1548-1568
[2]  
[Anonymous], 2006, Computer Vision and Pattern Recognition, 2006 IEEE Computer Society Conference on, DOI [DOI 10.1109/CVPR.2006.119, 10.1109/CVPR.2006.119]
[3]   Active appearance models [J].
Cootes, TF ;
Edwards, GJ ;
Taylor, CJ .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2001, 23 (06) :681-685
[4]   Histograms of oriented gradients for human detection [J].
Dalal, N ;
Triggs, B .
2005 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 1, PROCEEDINGS, 2005, :886-893
[5]   A spatio-temporal RBM-based model for facial expression recognition [J].
Elaiwat, S. ;
Bennamoun, M. ;
Boussaid, F. .
PATTERN RECOGNITION, 2016, 49 :152-161
[6]  
Fasel B., 2016, PATTERN RECOGN, V36, P259
[7]  
Goodfellow I, 2016, ADAPT COMPUT MACH LE, P1
[8]  
GOYANI M, 2017, INDIAN J SCI TECHNOL, V10, pNI147, DOI DOI 10.17485/ijst/2017/v10i9/108944
[9]   Facial expression recognition using radial encoding of local Gabor features and classifier synthesis [J].
Gu, Wenfei ;
Xiang, Cheng ;
Venkatesh, Y. V. ;
Huang, Dong ;
Lin, Hai .
PATTERN RECOGNITION, 2012, 45 (01) :80-91
[10]   Automatic Facial Expression Recognition Using Features of Salient Facial Patches [J].
Happy, S. L. ;
Routray, Aurobinda .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2015, 6 (01) :1-12