DDGAN: Dense Residual Module and Dual-stream Attention-Guided Generative Adversarial Network for colorizing near-infrared images

被引:8
作者
Chen, Yu [1 ]
Zhan, Weida [1 ]
Jiang, Yichun [1 ]
Zhu, Depeng [1 ]
Guo, Renzhong [1 ]
Xu, Xiaoyu [1 ]
机构
[1] Changchun Univ Sci, Technol Natl Demonstrat Ctr Expt Elect, Changchun 130022, Jilin, Peoples R China
关键词
Deep learning; Generative adversarial network; Near-infrared image colorization; Attention module;
D O I
10.1016/j.infrared.2023.104822
中图分类号
TH7 [仪器、仪表];
学科分类号
0804 ; 080401 ; 081102 ;
摘要
Transforming near-infrared(NIR) images into realistic RGB images is a challenging task. Recently, with the development of deep learning, the colorization of NIR images has been significantly improved. However, there are still problems with distorted textures and blurred details. The main reason is that NIR images require color prediction and gray-scale detail recovery to be colorized. Moreover, NIR images contain limited detail and lack color information, which poses a serious challenge for feature extraction. To address the problems, we propose the Dense Residual Module and Dual-stream Attention-Guided Generative Adversarial Network (DDGAN) in this paper. The Dense Residual Module (DRM) improves the network feature extraction capability in the form of dense residuals and increases the network depth to improve the adaptability of the network. The Dual-stream Attention Module (DAM) further improves the quality of colorized images by enhancing important features and suppressing unnecessary features to focus on essential visual features. We propose a composite loss function consisting of content loss, adversarial loss, perceptual loss, synthesized loss, and total variation loss, which improves the quality of colorful images in terms of both edge structure and visual perception. We consider the NIR images for the visible image conversion task to evaluate the efficiency and performance of the proposed model. The proposed DDGAN outperforms most existing methods in terms of efficiency and quality of the generated images on the RGB-NIR dataset and the OMSIV dataset. Compared to state-of-the-art methods, the proposed DDGAN shows promising results with significant improvements in PSNR, SSIM, MSE, and NRMSE. Extensive experimental data show that the proposed DDGAN can produce state-of-the-art colorized NIR images in objective metrics and subjective quality.
引用
收藏
页数:13
相关论文
共 56 条
[1]   Unsupervised object detection via LWIR/RGB translation [J].
Abbott, Rachael ;
Robertson, Neil M. ;
del Rincon, Jesus Martinez ;
Connor, Barry .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, :407-415
[2]   Image up-sampling using total-variation regularization with a new observation model [J].
Aly, HA ;
Dubois, E .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2005, 14 (10) :1647-1659
[3]   GSA-GAN: Global Spatial Attention Generative Adversarial Networks [J].
An, Lei ;
Zhao, Jiajia ;
Ma, Bo .
NEUROCOMPUTING, 2021, 437 :274-281
[4]   Recycle-GAN: Unsupervised Video Retargeting [J].
Bansal, Aayush ;
Ma, Shugao ;
Ramanan, Deva ;
Sheikh, Yaser .
COMPUTER VISION - ECCV 2018, PT V, 2018, 11209 :122-138
[5]   Generating Visible Spectrum Images from Thermal Infrared [J].
Berg, Amanda ;
Ahlberg, Jorgen ;
Felsberg, Michael .
PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2018, :1224-1233
[6]  
Bin J., 2021, Generative Adversarial Networks for Image-to-Image Translation, P313
[7]   Unsupervised Diverse Colorization via Generative Adversarial Networks [J].
Cao, Yun ;
Zhou, Zhiming ;
Zhang, Weinan ;
Yu, Yong .
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2017, PT I, 2017, 10534 :151-166
[8]   Colorization of infrared images based on feature fusion and contrastive learning [J].
Chen, Lingqiang ;
Liu, Yuan ;
He, Yin ;
Xie, Zhihua ;
Sui, Xiubao .
OPTICS AND LASERS IN ENGINEERING, 2023, 162
[9]   Beyond triplet loss: a deep quadruplet network for person re-identification [J].
Chen, Weihua ;
Chen, Xiaotang ;
Zhang, Jianguo ;
Huang, Kaiqi .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1320-1329
[10]   From Coarse to Fine (FC2F): A New Scheme of Colorizing Thermal Infrared Images [J].
Cheng, Feiyan ;
Shi, Junsheng ;
Yun, Lijun ;
Cao, Xuebing ;
Zhang, Jun .
IEEE ACCESS, 2020, 8 :111159-111171