Generating Face Images With Attributes for Free

被引:5
作者
Liu, Yaoyao [1 ,2 ]
Sun, Qianru [3 ]
He, Xiangnan [4 ]
Liu, An-An [1 ]
Su, Yuting [1 ]
Chua, Tat-Seng [5 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin, Peoples R China
[2] Max Planck Inst Informat, Saarland Informat Campus, Saarbrucken, Germany
[3] Singapore Management Univ, Sch Informat Syst, Singapore, Singapore
[4] Univ Sci & Technol China, Sch Informat Sci & Technol, Hefei, Peoples R China
[5] Natl Univ Singapore, Sch Comp, Singapore 119077, Singapore
基金
新加坡国家研究基金会; 中国国家自然科学基金;
关键词
Face; Face recognition; Image recognition; Image reconstruction; Task analysis; Gallium nitride; Decoding; Data augmentation; face attribute recognition; image generation; learning systems; pattern recognition;
D O I
10.1109/TNNLS.2020.3007790
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With superhuman-level performance of face recognition, we are more concerned about the recognition of fine-grained attributes, such as emotion, age, and gender. However, given that the label space is extremely large and follows a long-tail distribution, it is quite expensive to collect sufficient samples for fine-grained attributes. This results in imbalanced training samples and inferior attribute recognition models. To this end, we propose the use of arbitrary attribute combinations, without human effort, to synthesize face images. In particular, to bridge the semantic gap between high-level attribute label space and low-level face image, we propose a novel neural-network-based approach that maps the target attribute labels to an embedding vector, which can be fed into a pretrained image decoder to synthesize a new face image. Furthermore, to regularize the attribute for image synthesis, we propose to use a perceptual loss to make the new image explicitly faithful to target attributes. Experimental results show that our approach can generate photorealistic face images from attribute labels, and more importantly, by serving as augmented training samples, these images can significantly boost the performance of attribute recognition model. The code is open-sourced at this link.
引用
收藏
页码:2733 / 2743
页数:11
相关论文
共 51 条
  • [1] [Anonymous], 2010, ICML
  • [2] POOF: Part-Based One-vs-One Features for Fine-Grained Categorization, Face Verification, and Attribute Estimation
    Berg, Thomas
    Belhumeur, Peter N.
    [J]. 2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, : 955 - 962
  • [3] Data Augmentation-Based Joint Learning for Heterogeneous Face Recognition
    Cao, Bing
    Wang, Nannan
    Li, Jie
    Gao, Xinbo
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (06) : 1731 - 1743
  • [4] StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation
    Choi, Yunjey
    Choi, Minje
    Kim, Munyoung
    Ha, Jung-Woo
    Kim, Sunghun
    Choo, Jaegul
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 8789 - 8797
  • [5] Synthesizing Normalized Faces from Facial Identity Features
    Cole, Forrester
    Belanger, David
    Krishnan, Dilip
    Sarna, Aaron
    Mosseri, Inbar
    Freeman, William T.
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3386 - 3395
  • [6] Inverting the Generator of a Generative Adversarial Network
    Creswell, Antonia
    Bharath, Anil Anthony
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (07) : 1967 - 1974
  • [7] DOLL\AR P., 2007, Proc. Advances in Neural Information Processing Systems, V19, P361
  • [8] Creativity Inspired Zero-Shot Learning
    Elhoseiny, Mohamed
    Elfeki, Mohamed
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 5783 - 5792
  • [9] Goodfellow I., 2020, ADV NEUR IN, V63, P139, DOI [DOI 10.1145/3422622, 10.1145/3422622]
  • [10] AttGAN: Facial Attribute Editing by Only Changing What You Want
    He, Zhenliang
    Zuo, Wangmeng
    Kan, Meina
    Shan, Shiguang
    Chen, Xilin
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (11) : 5464 - 5478