Light the Way: An Enhanced Generative Adversarial Network Framework for Night-to-Day Image Translation With Improved Quality

被引:0
作者
Lakmal, H. K. I. S. [1 ]
Dissanayake, Maheshi B. [1 ]
Aramvith, Supavadee [2 ]
机构
[1] Univ Peradeniya, Fac Engn, Dept Elect & Elect Engn, Peradeniya 20400, Sri Lanka
[2] Chulalongkorn Univ, Fac Engn, Dept Elect Engn, Multimedia Data Analyt & Proc Res Unit, Bangkok 10330, Thailand
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Training; Image quality; Noise measurement; Generators; Brightness; Visualization; Roads; Benchmark testing; Accuracy; Vehicle safety; Generative adversarial network (GAN); image translation; night-time driving; night to day translation; perceptual loss;
D O I
10.1109/ACCESS.2024.3491792
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Driving at night introduces considerable challenges due to reduced visibility, making it essential to explore techniques that enhance road information for drivers. With this purview, the research presents a technique to address visibility constraints faced during night-time driving, by converting night-time road images to day-time images using a supervised Generative Adversarial Network (GAN) model (NtD-GAN). Since paired images are required to train supervised GAN models, the research first exploits a novel approach for generating paired night-day datasets, as it is practically infeasible to collect such image pairs in a natural setting, owing to dynamic traffic environments. An innovative generator network architecture is proposed for the NtD-GAN. Furthermore, a new approach was proposed for generating and loading initial weights to expedite the NtD-GAN training. This initial weight assignment resulted in faster convergence of the NtD-GAN with significant improvement in Inception Score (IS) by 17.3%, in Structural Similarity Index (SSIM) by 5.5%, and in Naturalness Image Quality Evaluator (NIQE) by 10.3%. Moreover, the perceptual loss is introduced to the training loss function of the NtD-GAN to increase the visual quality of the reconstructed images. The experimental results also demonstrated a 0.23% increment in IS, a 0.07% reduction in Fr & eacute;chet Inception Distance (FID), a 2.2% increment in SSIM, and a 7% reduction in Blind Referenceless Image Spatial Quality Evaluator (BRISQUE) compared to the NtD-GAN trained without perceptual loss. The comparison analysis with the benchmark models has demonstrated a significant improvement. For instance, in comparison to N2D-GAN, NtD-GAN has demonstrated a reduction in FID by 14.6%, an improvement in SSIM by 3.4%, an improvement in Peak Signal-to-Noise Ratio (PSNR) by 1.39 dB and a reduction in BRISQUE by 0.8%. The implementation of the NtD-GAN model is available at https://github.com/isurushanaka/paired-N2D.
引用
收藏
页码:165963 / 165978
页数:16
相关论文
共 92 条
  • [1] Almahairi A, 2018, Arxiv, DOI arXiv:1802.10151
  • [2] Anoosheh A, 2019, IEEE INT CONF ROBOT, P5958, DOI [10.1109/ICRA.2019.8794387, 10.1109/icra.2019.8794387]
  • [3] Ardiyanto I., 2021, Deep Learning and Big Data for Intelligent Transportation, P119
  • [4] Cross-Domain Car Detection Using Unsupervised Image-to-Image Translation: From Day to Night
    Arruda, Vinicius F.
    Paixao, Thiago M.
    Berriel, Rodrigo F.
    De Souza, Alberto F.
    Badue, Claudine
    Sebe, Nicu
    Oliveira-Santos, Thiago
    [J]. 2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [5] Benaim S, 2017, ADV NEUR IN, V30
  • [6] DUNIT: Detection-based Unsupervised Image-to-Image Translation
    Bhattacharjee, Deblina
    Kim, Seungryong
    Vizier, Guillaume
    Salzmann, Mathieu
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 4786 - 4795
  • [7] Cao JM, 2018, Arxiv, DOI arXiv:1805.08019
  • [8] Chang CC, 2009, PROCEEDINGS OF INTERNATIONAL SYMPOSIUM ON COMPUTER SCIENCE AND COMPUTATIONAL TECHNOLOGY (ISCSCT 2009), P1
  • [9] Chen RF, 2020, PROC CVPR IEEE, P8165, DOI 10.1109/CVPR42600.2020.00819
  • [10] The Cityscapes Dataset for Semantic Urban Scene Understanding
    Cordts, Marius
    Omran, Mohamed
    Ramos, Sebastian
    Rehfeld, Timo
    Enzweiler, Markus
    Benenson, Rodrigo
    Franke, Uwe
    Roth, Stefan
    Schiele, Bernt
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 3213 - 3223