E2F-Net: Eyes-to-face inpainting via StyleGAN latent space

被引:6
作者
Hassanpour, Ahmad [1 ]
Jamalbafrani, Fatemeh [2 ]
Yang, Bian [1 ]
Raja, Kiran [4 ]
Veldhuis, Raymond [1 ]
Fierrez, Julian [3 ]
机构
[1] NTNU, Dept Informat Secur & Commun Technol, Gjovik, Norway
[2] Sharif Univ Technol, Dept Elect Engn, Tehran, Iran
[3] Univ Autonoma Madrid, Sch Engn, Madrid, Spain
[4] NTNU, Dept Comp Sci, Gjovik, Norway
关键词
Eyes-to-face; Face inpainting; Face reconstruction; GAN latent space; StyleGAN; GENERATIVE ADVERSARIAL NETWORKS; IMAGE;
D O I
10.1016/j.patcog.2024.110442
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Face inpainting, the technique of restoring missing or damaged regions in facial images, is pivotal for applications like face recognition in occluded scenarios and image analysis with poor -quality captures. This process not only needs to produce realistic visuals but also preserve individual identity characteristics. The aim of this paper is to inpaint a face given periocular region (eyes -to -face) through a proposed new Generative Adversarial Network (GAN)-based model called Eyes -to -Face Network (E2F-Net). The proposed approach extracts identity and non-identity features from the periocular region using two dedicated encoders have been used. The extracted features are then mapped to the latent space of a pre -trained StyleGAN generator to benefit from its state-of-theart performance and its rich, diverse and expressive latent space without any additional training. We further improve the StyleGAN's output to find the optimal code in the latent space using a new optimization for GAN inversion technique. Our E2F-Net requires a minimum training process reducing the computational complexity as a secondary benefit. Through extensive experiments, we show that our method successfully reconstructs the whole face with high quality, surpassing current techniques, despite significantly less training and supervision efforts. We have generated seven eyes -to -face datasets based on well-known public face datasets for training and verifying our proposed methods. The code and datasets are publicly available.1
引用
收藏
页数:16
相关论文
共 54 条
[21]   Perceptual Losses for Real-Time Style Transfer and Super-Resolution [J].
Johnson, Justin ;
Alahi, Alexandre ;
Li Fei-Fei .
COMPUTER VISION - ECCV 2016, PT II, 2016, 9906 :694-711
[22]   A Style-Based Generator Architecture for Generative Adversarial Networks [J].
Karras, Tero ;
Laine, Samuli ;
Aila, Timo .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (12) :4217-4228
[23]  
Karras T, 2018, Arxiv, DOI arXiv:1710.10196
[24]   The contribution of different face parts to deep face recognition [J].
Lestriandoko, Nova Hadi ;
Veldhuis, Raymond ;
Spreeuwers, Luuk .
FRONTIERS IN COMPUTER SCIENCE, 2022, 4
[25]  
Li Y, 2021, P IEEE INT C IM PROC
[26]   Coherent Semantic Attention for Image Inpainting [J].
Liu, Hongyu ;
Jiang, Bin ;
Xiao, Yi ;
Yang, Chao .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :4169-4178
[27]   EyesGAN: Synthesize human face from human eyes [J].
Luo, Xiaodong ;
He, Xiaohai ;
Qing, Linbo ;
Chen, Xiang ;
Liu, Luping ;
Xu, Yining .
NEUROCOMPUTING, 2020, 404 :213-226
[28]  
Mahendran A, 2015, PROC CVPR IEEE, P5188, DOI 10.1109/CVPR.2015.7299155
[29]  
Mescheder L, 2018, PR MACH LEARN RES, V80
[30]  
Mori M., 1970, ENERGY, V7, P33, DOI DOI 10.1109/MRA.2012.2192811