Semi-Supervised Deep Neural Network for Joint Intensity Estimation of Multiple Facial Action Units

被引:1
作者
Zhang, Yong [1 ]
Fan, Yanbo [1 ]
Dong, Weiming [2 ]
Hu, Bao-Gang [2 ]
Ji, Qiang [3 ]
机构
[1] Tencent AI Lab, Shenzhen 518057, Guangdong, Peoples R China
[2] Chinese Acad Sci, Inst Automat, Natl Lab Pattern Recognit, Beijing 100190, Peoples R China
[3] Rensselaer Polytech Inst, Dept Elect Comp & Syst Engn, Troy, NY 12180 USA
基金
中国国家自然科学基金; 美国国家科学基金会;
关键词
Gold; Estimation; Hidden Markov models; Training; Face; Task analysis; Neural networks; Facial action units; intensity estimation; deep learning; weakly supervised learning; TRACKING; MODEL;
D O I
10.1109/ACCESS.2019.2947201
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Facial action units (AUs) are defined to depict movements of facial muscles, which are basic elements to encode facial expressions. Automatic AU intensity estimation is an important task in affective computing. Previous works leverage the representation power of deep neural networks (DNNs) to improve the performance of intensity estimation. However, a large number of intensity annotations are required to train DNNs that contain millions of parameters. But it is expensive and difficult to build a large-scale database with AU intensity annotation since AU annotation requires annotators have strong domain expertise. We propose a novel semi-supervised deep convolutional network that leverages extremely limited AU annotations for AU intensity estimation. It requires only intensity annotations of keyframes of training sequences. Domain knowledge on AUs is leveraged to provide weak supervisory information, including relative appearance similarity, temporal intensity ordering, facial symmetry, and contrastive appearance difference. We also propose a strategy to train a model for joint intensity estimation of multiple AUs under the setting of semi-supervised learning, which greatly improves the efficiency during inference. We perform empirical experiments on two public benchmark expression databases and make comparisons with state-of-the-art methods to demonstrate the effectiveness of the proposed method.
引用
收藏
页码:150743 / 150756
页数:14
相关论文
共 63 条
[1]  
[Anonymous], 2015, 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG)
[2]  
[Anonymous], IEEE T CYBERN
[3]  
[Anonymous], 2017, P IEEE INT C COMP VI
[4]  
[Anonymous], 2017, CVPR
[5]  
[Anonymous], 2016, PROC CVPR IEEE, DOI DOI 10.1109/CVPR.2016.532
[6]  
[Anonymous], 2017, IEEE C COMP VIS PATT
[7]  
[Anonymous], 2015, 2015 11 IEEE INT C W
[8]  
Baltruaitis T., 2015, 2015 11 IEEE INT C W, DOI [10.1109/FG.2015.7284869, DOI 10.1109/FG.2015.7284869]
[9]  
Baltrusaitis T, 2014, LECT NOTES COMPUT SC, V8692, P593, DOI 10.1007/978-3-319-10593-2_39
[10]   AUMPNet: simultaneous Action Units detection and intensity estimation on multipose facial images using a single convolutional neural network [J].
Batista, Julio Cesar ;
Albiero, Vitor ;
Bellon, Olga R. P. ;
Silva, Luciano .
2017 12TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2017), 2017, :866-871