PBR-GAN: Imitating Physically-Based Rendering With Generative Adversarial Networks

被引:1
作者
Li, Ru [1 ]
Dai, Peng [2 ]
Liu, Guanghui [3 ]
Zhang, Shengping [1 ]
Zeng, Bing [3 ]
Liu, Shuaicheng [3 ]
机构
[1] Harbin Inst Technol, Sch Comp Sci & Technol, Weihai 264209, Peoples R China
[2] Univ Hong Kong, Dept Elect & Elect Engn, Hong Kong, Peoples R China
[3] Univ Elect Sci & Technol China, Sch Informat & Commun Engn, Chengdu 611731, Peoples R China
基金
中国国家自然科学基金;
关键词
Rendering (computer graphics); Lighting; Decoding; Task analysis; Generative adversarial networks; Reflectivity; Color; Physically based rendering; generative adversarial network; illumination variation;
D O I
10.1109/TCSVT.2023.3298929
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
We propose a Generative Adversarial Network (GAN)-based architecture for achieving high-quality physically based rendering (PBR). Conventional PBR relies heavily on ray tracing, which is computationally expensive in complicated environments. Some recent deep learning-based methods can improve efficiency but cannot deal with illumination variation well. In this paper, we propose PBR-GAN, an end-to-end GAN-based network that solves these problems while generating natural photo-realistic images. Two encoders (the shading encoder and albedo encoder) and two decoders (the image decoder and light decoder) are introduced to achieve our target. The two encoders and the image decoder constitute the generator that learns the mapping between the generated domain and the real domain. The light decoder produces light maps that pay more attention to the highlight and shadow regions. The discriminator aims to optimize the generator by distinguishing target images from the generated ones. Three novel loss items, concentrating on domain translation, overall shading preservation, and light map estimation, are proposed to optimize the photo-realistic outputs. Furthermore, a real dataset is collected to provide realistic information for training GAN architecture. Extensive experiments indicate that PBR-GAN can preserve the illumination variation and improve the image perceptual quality.
引用
收藏
页码:1827 / 1840
页数:14
相关论文
共 58 条
[1]   SARGAN: Spatial Attention-Based Residuals for Facial Expression Manipulation [J].
Akram, Arbish ;
Khan, Nazar .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (10) :5433-5443
[2]  
[Anonymous], About us
[3]  
YOUTUBE
[4]  
baidu, About us
[5]   Mip-NeRF: A Multiscale Representation for Anti-Aliasing Neural Radiance Fields [J].
Barron, Jonathan T. ;
Mildenhall, Ben ;
Tancik, Matthew ;
Hedman, Peter ;
Martin-Brualla, Ricardo ;
Srinivasan, Pratul P. .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :5835-5844
[6]   Intrinsic Images in the Wild [J].
Bell, Sean ;
Bala, Kavita ;
Snavely, Noah .
ACM TRANSACTIONS ON GRAPHICS, 2014, 33 (04)
[7]  
Blender, About Us
[8]   A Simple Model for Intrinsic Image Decomposition with Depth Cues [J].
Chen, Qifeng ;
Koltun, Vladlen .
2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, :241-248
[9]   Learning Dynamic Generative Attention for Single Image Super-Resolution [J].
Chen, Rui ;
Zhang, Yan .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (12) :8368-8382
[10]   CartoonGAN: Generative Adversarial Networks for Photo Cartoonization [J].
Chen, Yang ;
Lai, Yu-Kun ;
Liu, Yong-Jin .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9465-9474