Semi-Supervised Deep Neural Network for Joint Intensity Estimation of Multiple Facial Action Units

被引:1
作者
Zhang, Yong [1 ]
Fan, Yanbo [1 ]
Dong, Weiming [2 ]
Hu, Bao-Gang [2 ]
Ji, Qiang [3 ]
机构
[1] Tencent AI Lab, Shenzhen 518057, Guangdong, Peoples R China
[2] Chinese Acad Sci, Inst Automat, Natl Lab Pattern Recognit, Beijing 100190, Peoples R China
[3] Rensselaer Polytech Inst, Dept Elect Comp & Syst Engn, Troy, NY 12180 USA
基金
中国国家自然科学基金; 美国国家科学基金会;
关键词
Gold; Estimation; Hidden Markov models; Training; Face; Task analysis; Neural networks; Facial action units; intensity estimation; deep learning; weakly supervised learning; TRACKING; MODEL;
D O I
10.1109/ACCESS.2019.2947201
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Facial action units (AUs) are defined to depict movements of facial muscles, which are basic elements to encode facial expressions. Automatic AU intensity estimation is an important task in affective computing. Previous works leverage the representation power of deep neural networks (DNNs) to improve the performance of intensity estimation. However, a large number of intensity annotations are required to train DNNs that contain millions of parameters. But it is expensive and difficult to build a large-scale database with AU intensity annotation since AU annotation requires annotators have strong domain expertise. We propose a novel semi-supervised deep convolutional network that leverages extremely limited AU annotations for AU intensity estimation. It requires only intensity annotations of keyframes of training sequences. Domain knowledge on AUs is leveraged to provide weak supervisory information, including relative appearance similarity, temporal intensity ordering, facial symmetry, and contrastive appearance difference. We also propose a strategy to train a model for joint intensity estimation of multiple AUs under the setting of semi-supervised learning, which greatly improves the efficiency during inference. We perform empirical experiments on two public benchmark expression databases and make comparisons with state-of-the-art methods to demonstrate the effectiveness of the proposed method.
引用
收藏
页码:150743 / 150756
页数:14
相关论文
共 63 条
[11]   Data-Free Prior Model for Upper Body Pose Estimation and Tracking [J].
Chen, Jixu ;
Nie, Siqi ;
Ji, Qiang .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2013, 22 (12) :4627-4639
[12]   FaceNet2ExpNet: Regularizing a Deep Face Recognition Net for Expression Recognition [J].
Ding, Hui ;
Zhou, Shaohua Kevin ;
Chellappa, Rama .
2017 12TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2017), 2017, :118-126
[13]  
Ekman P., 1978, APA PsycTests, DOI DOI 10.1037/T27734-000
[14]   Optimizing Filter Size in Convolutional Neural Networks for Facial Action Unit Recognition [J].
Han, Shizhong ;
Meng, Zibo ;
Li, Zhiyuan ;
O'Reilly, James ;
Cai, Jie ;
Wang, Xiaofeng ;
Tong, Yan .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :5070-5078
[15]  
Jyothirmaye S, 2019, P 14 IEEE INT C AUT, P1
[16]  
Kaltwang S, 2015, PROC CVPR IEEE, P296, DOI 10.1109/CVPR.2015.7298626
[17]  
Kaltwang S, 2012, LECT NOTES COMPUT SC, V7432, P368, DOI 10.1007/978-3-642-33191-6_36
[18]  
Kaltwang Sebastian, 2015, IEEE T PATTERN ANAL, P1748
[19]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[20]  
Lee Dong-Hyun, 2013, P WORKSH CHALL REPR