High Efficient Spatial and Radiation Information Mutual Enhancing Fusion Method for Visible and Infrared Image

被引:1
作者
Liu, Zongzhen [1 ,2 ,3 ,4 ,5 ]
Wei, Yuxing [2 ,3 ,4 ,5 ]
Huang, Geli [1 ]
Li, Chao [1 ]
Zhang, Jianlin [2 ,3 ,4 ,5 ]
Li, Meihui [2 ,3 ,4 ,5 ]
Liu, Dongxu [2 ,3 ,4 ,5 ]
Peng, Xiaoming [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Automat Engn, Chengdu 611731, Peoples R China
[2] Chinese Acad Sci, Natl Key Lab Opt Field Manipulat Sci & Technol, Chengdu 610209, Peoples R China
[3] Chinese Acad Sci, Key Lab Opt Engn, Chengdu 610209, Peoples R China
[4] Chinese Acad Sci, Inst Opt & Elect, Chengdu 610209, Peoples R China
[5] Univ Chinese Acad Sci, Beijing 101408, Peoples R China
关键词
Image fusion; cross-modal; light perception; visible and infrared image; mutual enhancement; NETWORK; ARCHITECTURE; ENHANCEMENT; JOINT; NEST;
D O I
10.1109/ACCESS.2024.3351774
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Visible and infrared image fusion is an important image enhancement technique that aims to generate high-quality fused images with prominent targets and rich textures in extreme environments. However, most of the current image fusion methods have poor visual perception of the generated fused images due to severe degradation of texture details of the visible light images in scenes with extreme lighting, which seriously affects the application of subsequent advanced vision tasks such as target detection and tracking. To address these challenges, this paper bridges the gap between image fusion and advanced vision tasks by proposing an efficient fusion method for mutual enhancement of spatial and radiometric information. First, we design the gradient residual dense block (LGCnet) to improve the description of fine spatial details in the fusion network. Then, we developed a cross-modal perceptual fusion (CMPF) module to facilitate modal interactions in the network, which effectively enhances the fusion of complementary information between different modalities and reduces redundant learning. Finally, we designed an adaptive light-aware network (ALPnet) to guide the training of the fusion network to facilitate the fusion network to adaptively select more effective information for fusion under different lighting conditions. Extensive experiments show that the proposed fusion approach has competitive advantages over six current state-of-the-art deep-learning methods in highlighting target features and describing the global scene.
引用
收藏
页码:6971 / 6992
页数:22
相关论文
共 62 条
[11]   NestFuse: An Infrared and Visible Image Fusion Architecture Based on Nest Connection and Spatial/Channel Attention Models [J].
Li, Hui ;
Wu, Xiao-Jun ;
Durrani, Tariq .
IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2020, 69 (12) :9645-9656
[12]   Fast infrared and visible image fusion with structural decomposition [J].
Li, Hui ;
Qi, Xianbiao ;
Xie, Wuyuan .
KNOWLEDGE-BASED SYSTEMS, 2020, 204
[13]   MDLatLRR: A Novel Decomposition Method for Infrared and Visible Image Fusion [J].
Li, Hui ;
Wu, Xiao-Jun ;
Kittler, Josef .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :4733-4746
[14]   DenseFuse: A Fusion Approach to Infrared and Visible Images [J].
Li, Hui ;
Wu, Xiao-Jun .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (05) :2614-2623
[15]   AttentionFGAN: Infrared and Visible Image Fusion Using Attention-Based Generative Adversarial Networks [J].
Li, Jing ;
Huo, Hongtao ;
Li, Chang ;
Wang, Renhua ;
Feng, Qi .
IEEE TRANSACTIONS ON MULTIMEDIA, 2021, 23 :1383-1396
[16]   Pixel-level image fusion: A survey of the state of the art [J].
Li, Shutao ;
Kang, Xudong ;
Fang, Leyuan ;
Hu, Jianwen ;
Yin, Haitao .
INFORMATION FUSION, 2017, 33 :100-112
[17]  
Liu G., 2010, P ICML, V1, P8
[18]   Robust Recovery of Subspace Structures by Low-Rank Representation [J].
Liu, Guangcan ;
Lin, Zhouchen ;
Yan, Shuicheng ;
Sun, Ju ;
Yu, Yong ;
Ma, Yi .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (01) :171-184
[19]   Target-aware Dual Adversarial Learning and a Multi-scenario Multi-Modality Benchmark to Fuse Infrared and Visible for Object Detection [J].
Liu, Jinyuan ;
Fan, Xin ;
Huang, Zhanbo ;
Wu, Guanyao ;
Liu, Risheng ;
Zhong, Wei ;
Luo, Zhongxuan .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :5792-5801
[20]   Learning a Deep Multi-Scale Feature Ensemble and an Edge-Attention Guidance for Image Fusion [J].
Liu, Jinyuan ;
Fan, Xin ;
Jiang, Ji ;
Liu, Risheng ;
Luo, Zhongxuan .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (01) :105-119