Hard negative generation for identity-disentangled facial expression recognition

被引:76
作者
Liu, Xiaofeng [1 ,2 ,3 ]
Kumar, B. V. K. Vijaya [3 ,4 ]
Jia, Ping [1 ,2 ]
You, Jane [1 ,2 ,5 ]
机构
[1] Chinese Acad Sci, Changchun Inst Opt Fine Mech & Phys, Changchun, Jilin, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
[3] Carnegie Mellon Univ, Dept Elect & Comp Engn, Pittsburgh, PA 15213 USA
[4] Carnegie Mellon Univ Africa, Kigali, Rwanda
[5] Hong Kong Polytech Univ, Dept Comp, Hong Kong, Peoples R China
关键词
Hard negative generation; Adaptive metric learning; Face normalization; Facial expression recognition;
D O I
10.1016/j.patcog.2018.11.001
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Various factors such as identity-specific attributes, pose, illumination and expression affect the appearance of face images. Disentangling the identity-specific factors is potentially beneficial for facial expression recognition (FER). Existing image-based FER systems either use hand-crafted or learned features to represent a single face image. In this paper, we propose a novel FER framework, named identity disentangled facial expression recognition machine (IDFERM), in which we untangle the identity from a query sample by exploiting its difference from its references (e.g., its mined or generated frontal and neutral normalized faces). We demonstrate a possible 'recognition via generation' scheme which consists of a novel hard negative generation (HNG) network and a generalized radial metric learning (RML) network. For FER, generated normalized faces are used as hard negative samples for metric learning. The difficulty of threshold validation and anchor selection are alleviated in RML and its distance comparisons are fewer than those of traditional deep metric learning methods. The expression representations of RML achieve superior performance on the CK +, MMI and Oulu-CASIA datasets, given a single query image for testing. (C) 2018 Elsevier Ltd. All rights reserved.
引用
收藏
页码:1 / 12
页数:12
相关论文
共 47 条
[11]   FaceNet2ExpNet: Regularizing a Deep Face Recognition Net for Expression Recognition [J].
Ding, Hui ;
Zhou, Shaohua Kevin ;
Chellappa, Rama .
2017 12TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2017), 2017, :118-126
[12]   Deep feature learning with relative distance comparison for person re-identification [J].
Ding, Shengyong ;
Lin, Liang ;
Wang, Guangrun ;
Chao, Hongyang .
PATTERN RECOGNITION, 2015, 48 (10) :2993-3003
[13]   LAM3L: Locally adaptive maximum margin metric learning for visual data classification [J].
Dong, Yanni ;
Du, Bo ;
Zhang, Lefei ;
Zhang, Liangpei ;
Tao, Dacheng .
NEUROCOMPUTING, 2017, 235 :1-9
[14]  
Goodfellow I., 2016, ARXIV
[15]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672
[16]   Multi-PIE [J].
Gross, Ralph ;
Matthews, Iain ;
Cohn, Jeffrey ;
Kanade, Takeo ;
Baker, Simon .
IMAGE AND VISION COMPUTING, 2010, 28 (05) :807-813
[17]  
Han S., 2016, ADV NEURAL INF PROCE, P109, DOI 10.5555/3157096.3157109
[18]   The distributed human neural system for face perception [J].
Haxby, JV ;
Hoffman, EA ;
Gobbini, MI .
TRENDS IN COGNITIVE SCIENCES, 2000, 4 (06) :223-233
[19]  
He K., 2015, IEEE I CONF COMP VIS, P1026, DOI DOI 10.1109/ICCV.2015.123
[20]   Joint Fine-Tuning in Deep Neural Networks for Facial Expression Recognition [J].
Jung, Heechul ;
Lee, Sihaeng ;
Yim, Junho ;
Park, Sunjeong ;
Kim, Junmo .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :2983-2991