Cross-Domain and Disentangled Face Manipulation With 3D Guidance

被引:7
作者
Wang, Can [1 ]
Chai, Menglei [2 ]
He, Mingming [3 ]
Chen, Dongdong [4 ]
Liao, Jing [1 ]
机构
[1] City Univ Hong Kong, Dept Comp Sci, Hong Kong 999077, Peoples R China
[2] Snap Inc, Creat Vis Team, Santa Monica, CA 90405 USA
[3] Univ Southern Calif, Inst Creat Technol, Los Angeles, CA 90007 USA
[4] Microsoft Cloud AI, Redmond, WA 98052 USA
关键词
Faces; Aerospace electronics; Three-dimensional displays; Semantics; Lighting; Solid modeling; Codes; Face image manipulation; domain adaptation; 3D morphable model; disentanglement; StyleGAN2; RECONSTRUCTION; IMAGE; VIDEO;
D O I
10.1109/TVCG.2021.3139913
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Face image manipulation via three-dimensional guidance has been widely applied in various interactive scenarios due to its semantically-meaningful understanding and user-friendly controllability. However, existing 3D-morphable-model-based manipulation methods are not directly applicable to out-of-domain faces, such as non-photorealistic paintings, cartoon portraits, or even animals, mainly due to the formidable difficulties in building the model for each specific face domain. To overcome this challenge, we propose, as far as we know, the first method to manipulate faces in arbitrary domains using human 3DMM. This is achieved through two major steps: 1) disentangled mapping from 3DMM parameters to the latent space embedding of a pre-trained StyleGAN2 [1] that guarantees disentangled and precise controls for each semantic attribute; and 2) cross-domain adaptation that bridges domain discrepancies and makes human 3DMM applicable to out-of-domain faces by enforcing a consistent latent space embedding. Experiments and comparisons demonstrate the superiority of our high-quality semantic manipulation method on a variety of face domains with all major 3D facial attributes controllable - pose, expression, shape, albedo, and illumination. Moreover, we develop an intuitive editing interface to support user-friendly control and instant feedback. Our project page is https://cassiepython.github.io/cddfm3d/index.html.
引用
收藏
页码:2053 / 2066
页数:14
相关论文
共 59 条
[1]   StyleFlow: Attribute-conditioned Exploration of StyleGAN-Generated Images using Conditional Continuous Normalizing Flows [J].
Abdal, Rameen ;
Zhu, Peihao ;
Mitra, Niloy J. ;
Wonka, Peter .
ACM TRANSACTIONS ON GRAPHICS, 2021, 40 (03)
[2]   Image2StyleGAN: How to Embed Images Into the StyleGAN Latent Space? [J].
Abdal, Rameen ;
Qin, Yipeng ;
Wonka, Peter .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :4431-4440
[3]  
[Anonymous], 2021, Danbooru2020: A large-scale crowdsourced and tagged anime illustration dataset
[4]   Facial Performance Transfer via Deformable Models and Parametric Correspondence [J].
Asthana, Akshay ;
Delahunty, Miles ;
Dhall, Abhinav ;
Goecke, Roland .
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2012, 18 (09) :1511-1519
[5]   Bringing Portraits to Life [J].
Averbuch-Elor, Hadar ;
Cohen-Or, Daniel ;
Kopf, Johannes ;
Cohen, Michael F. .
ACM TRANSACTIONS ON GRAPHICS, 2017, 36 (06)
[6]   A morphable model for the synthesis of 3D faces [J].
Blanz, V ;
Vetter, T .
SIGGRAPH 99 CONFERENCE PROCEEDINGS, 1999, :187-194
[7]   FaceWarehouse: A 3D Facial Expression Database for Visual Computing [J].
Cao, Chen ;
Weng, Yanlin ;
Zhou, Shun ;
Tong, Yiying ;
Zhou, Kun .
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2014, 20 (03) :413-425
[8]   Coherent Online Video Style Transfer [J].
Chen, Dongdong ;
Liao, Jing ;
Yuan, Lu ;
Yu, Nenghai ;
Hua, Gang .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :1114-1123
[9]   StarGAN v2: Diverse Image Synthesis for Multiple Domains [J].
Choi, Yunjey ;
Uh, Youngjung ;
Yoo, Jaejun ;
Ha, Jung-Woo .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :8185-8194
[10]   Video Face Replacement [J].
Dale, Kevin ;
Sunkavalli, Kalyan ;
Johnson, Micah K. ;
Vlasic, Daniel ;
Matusik, Wojciech ;
Pfister, Hanspeter .
ACM TRANSACTIONS ON GRAPHICS, 2011, 30 (06)