Facial Expression Recognition Methods in the Wild Based on Fusion Feature of Attention Mechanism and LBP

被引:17
作者
Liao, Jun [1 ,2 ,3 ,4 ]
Lin, Yuanchang [1 ,3 ,4 ]
Ma, Tengyun [1 ,3 ,4 ]
He, Songxiying [3 ,4 ]
Liu, Xiaofang [1 ,3 ,4 ]
He, Guotian [1 ,3 ,4 ]
机构
[1] Chinese Acad Sci, Chongqing Inst Green Intelligent Technol, Chongqing 400714, Peoples R China
[2] Chongqing Univ Technol, Coll Mech Engn, Chongqing 400054, Peoples R China
[3] Chinese Acad Sci, Chongqing Inst Green Intelligent Technol, Chongqing Key Lab Artificial Intelligence, Chongqing 400714, Peoples R China
[4] Chinese Acad Sci, Chongqing Inst Green Intelligent Technol, Serv Robot Control Technol, Chongqing 400714, Peoples R China
关键词
deep learning; facial expression recognition; attention mechanism; LBP features; MODELS; SCALE; GAN;
D O I
10.3390/s23094204
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Facial expression methods play a vital role in human-computer interaction and other fields, but there are factors such as occlusion, illumination, and pose changes in wild facial recognition, as well as category imbalances between different datasets, that result in large variations in recognition rates and low accuracy rates for different categories of facial expression datasets. This study introduces RCL-Net, a method of recognizing wild facial expressions that is based on an attention mechanism and LBP feature fusion. The structure consists of two main branches, namely the ResNet-CBAM residual attention branch and the local binary feature (LBP) extraction branch (RCL-Net). First, by merging the residual network and hybrid attention mechanism, the residual attention network is presented to emphasize the local detail feature information of facial expressions; the significant characteristics of facial expressions are retrieved from both channel and spatial dimensions to build the residual attention classification model. Second, we present a locally improved residual network attention model. LBP features are introduced into the facial expression feature extraction stage in order to extract texture information on expression photographs in order to emphasize facial feature information and enhance the recognition accuracy of the model. Lastly, experimental validation is performed using the FER2013, FERPLUS, CK+, and RAF-DB datasets, and the experimental results demonstrate that the proposed method has superior generalization capability and robustness in the laboratory-controlled environment and field environment compared to the most recent experimental methods.
引用
收藏
页数:19
相关论文
共 57 条
[1]   Identity preserving multi-pose facial expression recognition using fine tuned VGG on the latent space vector of generative adversarial network [J].
Abiram, R. Nandhini ;
Vincent, P. M. Durai Raj .
MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2021, 18 (04) :3699-3717
[2]   Emotion Recognition in Speech using Cross-Modal Transfer in the Wild [J].
Albanie, Samuel ;
Nagrani, Arsha ;
Vedaldi, Andrea ;
Zisserman, Andrew .
PROCEEDINGS OF THE 2018 ACM MULTIMEDIA CONFERENCE (MM'18), 2018, :292-301
[3]   Deep Pain: Exploiting Long Short-Term Memory Networks for Facial Expression Classification [J].
Rodriguez, Pau ;
Cucurull, Guillem ;
Gonzalez, Jordi ;
Gonfaus, Josep M. ;
Nasrollahi, Kamal ;
Moeslund, Thomas B. ;
Roca, F. Xavier .
IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (05) :3314-3324
[4]   Training Deep Networks for Facial Expression Recognition with Crowd-Sourced Label Distribution [J].
Barsoum, Emad ;
Zhang, Cha ;
Ferrer, Cristian Canton ;
Zhang, Zhengyou .
ICMI'16: PROCEEDINGS OF THE 18TH ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2016, :279-283
[5]   Recognizing facial actions using gabor wavelets with neutral face average difference [J].
Bazzo, JJ ;
Lamar, MV .
SIXTH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, PROCEEDINGS, 2004, :505-510
[6]   EmotioNet: An accurate, real-time algorithm for the automatic annotation of a million facial expressions in the wild [J].
Benitez-Quiroz, C. Fabian ;
Srinivasan, Ramprakash ;
Martinez, Aleix M. .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :5562-5570
[7]  
Chang TY, 2018, Arxiv, DOI arXiv:1803.00185
[8]  
Cheng CL, 2017, 2017 12TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS AND KNOWLEDGE ENGINEERING (IEEE ISKE)
[9]   Histograms of oriented gradients for human detection [J].
Dalal, N ;
Triggs, B .
2005 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 1, PROCEEDINGS, 2005, :886-893
[10]  
Dhall A, 2017, PROCEEDINGS OF THE 19TH ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, ICMI 2017, P524, DOI 10.1145/3136755.3143004