EnlightenGAN: Deep Light Enhancement Without Paired Supervision

被引:1318
作者
Jiang, Yifan [1 ]
Gong, Xinyu [1 ]
Liu, Ding [2 ]
Cheng, Yu [2 ,3 ]
Fang, Chen [2 ]
Shen, Xiaohui [2 ]
Yang, Jianchao [2 ]
Zhou, Pan [4 ]
Wang, Zhangyang [1 ]
机构
[1] Univ Texas Austin, Dept Elect & Comp Engn, Austin, TX 78712 USA
[2] Bytedance Inc, Mountain View, CA 94041 USA
[3] Microsoft AI & Res, Redmond, WA 98052 USA
[4] Huazhong Univ Sci & Technol, Dept Elect Informat & Commun, Wuhan 430074, Peoples R China
关键词
Training; Visualization; Lighting; Generative adversarial networks; Gallium nitride; Adaptation models; Training data; Low-light enhancement; generative adversarial networks; unsupervised learning; RETINEX;
D O I
10.1109/TIP.2021.3051462
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning-based methods have achieved remarkable success in image restoration and enhancement, but are they still competitive when there is a lack of paired training data? As one such example, this paper explores the low-light image enhancement problem, where in practice it is extremely challenging to simultaneously take a low-light and a normal-light photo of the same visual scene. We propose a highly effective unsupervised generative adversarial network, dubbed EnlightenGAN, that can be trained without low/normal-light image pairs, yet proves to generalize very well on various real-world test images. Instead of supervising the learning using ground truth data, we propose to regularize the unpaired training using the information extracted from the input itself, and benchmark a series of innovations for the low-light image enhancement problem, including a global-local discriminator structure, a self-regularized perceptual loss fusion, and the attention mechanism. Through extensive experiments, our proposed approach outperforms recent methods under a variety of metrics in terms of visual quality and subjective user study. Thanks to the great flexibility brought by unpaired training, EnlightenGAN is demonstrated to be easily adaptable to enhancing real-world images from various domains. Our codes and pre-trained models are available at: https://github.com/VITA-Group/EnlightenGAN.
引用
收藏
页码:2340 / 2349
页数:10
相关论文
共 58 条
[1]  
[Anonymous], 2018, P EUR C COMP VIS ECC
[2]  
BRADLEY RA, 1952, BIOMETRIKA, V39, P324, DOI 10.1093/biomet/39.3-4.324
[3]   Learning a Deep Single Image Contrast Enhancer from Multi-Exposure Images [J].
Cai, Jianrui ;
Gu, Shuhang ;
Zhang, Lei .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (04) :2049-2062
[4]   Learning to See in the Dark [J].
Chen, Chen ;
Chen, Qifeng ;
Xu, Jia ;
Koltun, Vladlen .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :3291-3300
[5]   Deep Photo Enhancer: Unpaired Learning for Image Enhancement from Photographs with GANs [J].
Chen, Yu-Sheng ;
Wang, Yu-Ching ;
Kao, Man-Hsin ;
Chuang, Yung-Yu .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :6306-6314
[6]   StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation [J].
Choi, Yunjey ;
Choi, Minje ;
Kim, Munyoung ;
Ha, Jung-Woo ;
Kim, Sunghun ;
Choo, Jaegul .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8789-8797
[7]  
Dang-Nguyen DT., 2015, P 6 ACM MULT SYST C, P219
[8]   A weighted variational model for simultaneous reflectance and illumination estimation [J].
Fu, Xueyang ;
Zeng, Delu ;
Huang, Yue ;
Zhang, Xiao-Ping ;
Ding, Xinghao .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :2782-2790
[9]   Deep Bilateral Learning for Real-Time Image Enhancement [J].
Gharbi, Michael ;
Chen, Jiawen ;
Barron, Jonathan T. ;
Hasinoff, Samuel W. ;
Durand, Fredo .
ACM TRANSACTIONS ON GRAPHICS, 2017, 36 (04)
[10]   AutoGAN: Neural Architecture Search for Generative Adversarial Networks [J].
Gong, Xinyu ;
Chang, Shiyu ;
Jiang, Yifan ;
Wang, Zhangyang .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :3223-3233