Controllable and Identity-Aware Facial Attribute Transformation

被引:13
作者
Tan, Daniel Stanley [1 ]
Soeseno, Jonathan Hans [1 ]
Hua, Kai-Lung [1 ]
机构
[1] Natl Taiwan Univ Sci & Technol, CSIE, Taipei 106, Taiwan
关键词
Facial features; Generators; Face recognition; Faces; Hair; Image color analysis; Generative adversarial networks; Controllable transformation; facial attributes; generative adversarial network (GAN); multitask discriminator; self-cycle loss;
D O I
10.1109/TCYB.2021.3071172
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Modifying facial attributes without the paired dataset proves to be a challenging task. Previously, approaches either required supervision from a ground-truth transformed image or required training a separate model for mapping every pair of attributes. These limit the scalability of the models to accommodate a larger set of attributes since the number of models that we need to train grows exponentially large. Another major drawback of the previous approaches is the unintentional gain of the identity of the person as they transform the facial attributes. We propose a method that allows for controllable and identity-aware transformations across multiple facial attributes using only a single model. Our approach is to train a generative adversarial network (GAN) with a multitask conditional discriminator that recognizes the identity of the face, distinguishes real images from fake, as well as identifies facial attributes present in an image. This guides the generator into producing an output that is realistic while preserving the person's identity and facial attributes. Through this framework, our model also learns meaningful image representations in a lower dimensional latent space and semantically associate separate parts of the encoded vector with both the person's identity and facial attributes. This opens up the possibility of generating new faces and other transformations such as making the face thinner or chubbier. Furthermore, our model only encodes the image once and allows for multiple transformations using the encoded vector. This allows for faster transformations since it does not need to reprocess the entire image for every transformation. We show the effectiveness of our proposed method through both qualitative and quantitative evaluations, such as ablative studies, visual inspection, and face verification. Competitive results are achieved compared to the main competition (CycleGAN), however, at great space and extensibility gain by using a single model.
引用
收藏
页码:4825 / 4836
页数:12
相关论文
共 48 条
[1]  
Amos B, 2016, CMU SCH COMPUTER SCI, V6
[2]  
Antipov G, 2017, IEEE IMAGE PROC, P2089, DOI 10.1109/ICIP.2017.8296650
[3]  
Chen X, 2016, 30 C NEURAL INFORM P, V29
[4]   3D Object Completion via Class-Conditional Generative Adversarial Network [J].
Chen, Yu-Chieh ;
Tan, Daniel Stanley ;
Cheng, Wen-Huang ;
Hua, Kai-Lung .
MULTIMEDIA MODELING, MMM 2019, PT II, 2019, 11296 :54-66
[5]   StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation [J].
Choi, Yunjey ;
Choi, Minje ;
Kim, Munyoung ;
Ha, Jung-Woo ;
Kim, Sunghun ;
Choo, Jaegul .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8789-8797
[6]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672
[7]  
Guo D, 2009, PROC CVPR IEEE, P73, DOI 10.1109/CVPRW.2009.5206833
[8]   Stacked Generative Adversarial Networks [J].
Huang, Xun ;
Li, Yixuan ;
Poursaeed, Omid ;
Hopcroft, John ;
Belongie, Serge .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1866-1875
[9]   Image-to-Image Translation with Conditional Adversarial Networks [J].
Isola, Phillip ;
Zhu, Jun-Yan ;
Zhou, Tinghui ;
Efros, Alexei A. .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :5967-5976
[10]  
Kingma DP, 2014, ADV NEUR IN, V27