DDGAN: Dense Residual Module and Dual-stream Attention-Guided Generative Adversarial Network for colorizing near-infrared images

被引:8
作者
Chen, Yu [1 ]
Zhan, Weida [1 ]
Jiang, Yichun [1 ]
Zhu, Depeng [1 ]
Guo, Renzhong [1 ]
Xu, Xiaoyu [1 ]
机构
[1] Changchun Univ Sci, Technol Natl Demonstrat Ctr Expt Elect, Changchun 130022, Jilin, Peoples R China
关键词
Deep learning; Generative adversarial network; Near-infrared image colorization; Attention module;
D O I
10.1016/j.infrared.2023.104822
中图分类号
TH7 [仪器、仪表];
学科分类号
0804 ; 080401 ; 081102 ;
摘要
Transforming near-infrared(NIR) images into realistic RGB images is a challenging task. Recently, with the development of deep learning, the colorization of NIR images has been significantly improved. However, there are still problems with distorted textures and blurred details. The main reason is that NIR images require color prediction and gray-scale detail recovery to be colorized. Moreover, NIR images contain limited detail and lack color information, which poses a serious challenge for feature extraction. To address the problems, we propose the Dense Residual Module and Dual-stream Attention-Guided Generative Adversarial Network (DDGAN) in this paper. The Dense Residual Module (DRM) improves the network feature extraction capability in the form of dense residuals and increases the network depth to improve the adaptability of the network. The Dual-stream Attention Module (DAM) further improves the quality of colorized images by enhancing important features and suppressing unnecessary features to focus on essential visual features. We propose a composite loss function consisting of content loss, adversarial loss, perceptual loss, synthesized loss, and total variation loss, which improves the quality of colorful images in terms of both edge structure and visual perception. We consider the NIR images for the visible image conversion task to evaluate the efficiency and performance of the proposed model. The proposed DDGAN outperforms most existing methods in terms of efficiency and quality of the generated images on the RGB-NIR dataset and the OMSIV dataset. Compared to state-of-the-art methods, the proposed DDGAN shows promising results with significant improvements in PSNR, SSIM, MSE, and NRMSE. Extensive experimental data show that the proposed DDGAN can produce state-of-the-art colorized NIR images in objective metrics and subjective quality.
引用
收藏
页数:13
相关论文
共 56 条
[51]   Colorful Image Colorization [J].
Zhang, Richard ;
Isola, Phillip ;
Efros, Alexei A. .
COMPUTER VISION - ECCV 2016, PT III, 2016, 9907 :649-666
[52]   Occlusion-Aware R-CNN: Detecting Pedestrians in a Crowd [J].
Zhang, Shifeng ;
Wen, Longyin ;
Bian, Xiao ;
Lei, Zhen ;
Li, Stan Z. .
COMPUTER VISION - ECCV 2018, PT III, 2018, 11207 :657-674
[53]  
[赵源萌 Zhao Yuanmeng], 2009, [光学学报, Acta Optica Sinica], V29, P654
[54]   SCGAN: Saliency Map-Guided Colorization With Generative Adversarial Network [J].
Zhao, Yuzhi ;
Po, Lai-Man ;
Cheung, Kwok-Wai ;
Yu, Wing-Yin ;
Rehman, Yasar Abbas Ur .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (08) :3062-3077
[55]  
Zheng W.F., 2022, arXiv
[56]   Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks [J].
Zhu, Jun-Yan ;
Park, Taesung ;
Isola, Phillip ;
Efros, Alexei A. .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :2242-2251