A robust infrared and visible image fusion framework via multi-receptive-field attention and color visual perception

被引:19
作者
Ding, Zhaisheng [1 ]
Li, Haiyan [1 ]
Zhou, Dongming [1 ]
Liu, Yanyu [1 ]
Hou, Ruichao [2 ]
机构
[1] Yunnan Univ, Sch Informat & Artificial Intelligence, Kunming 650504, Yunnan, Peoples R China
[2] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing 210023, Peoples R China
基金
中国国家自然科学基金;
关键词
Infrared and visible image; Image fusion; Deep learning; Visual perception quality; QUALITY ASSESSMENT; NETWORK; NEST;
D O I
10.1007/s10489-022-03952-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, a robust infrared and visible image fusion scheme that joins a dual-branch multi-receptive-field neural network and a color vision transfer algorithm is designed to aggregate infrared and visible video sequences. The proposed method enables the fused image to effectively recognize thermal objects, contain rich texture information and ensure visual perception quality. The fusion network is an integrated encoder-decoder modal with a multi-receptive-field attention mechanism that is implemented via hybrid dilated convolution (HDC) and a series of convolution layers to form an unsupervised framework. Specifically, the multi-receptive-field attention mechanism aims to extract comprehensive spatial information to enable the encoder to separately focus on the substantial thermal radiation from the infrared modal and the environmental characteristics from the visible modal. In addition, to ensure that the fused image has rich color, high fidelity and steady brightness, a color vision transfer method is proposed to recolor the fused gray results by deriving a map from the visible image serving as a reference. Extensive experiments verify the importance and robustness of each step in the subjective and objective evaluation and demonstrate that our work represents a trade-off among color fidelity, fusion performance and computational efficiency. Moreover, we will publish our research content, data and code publicly at https://github.com/DZSYUNNAN/RGB-TIR-image-fusion.
引用
收藏
页码:8114 / 8132
页数:19
相关论文
共 41 条
[1]  
A-Monem ME, 2020, IRAQI J SCI, P675
[2]  
Alexander T., 2014, TNO Image Fusion Dataset
[3]  
[Anonymous], INO VIDEO DATASET
[4]  
[Anonymous], 2012, SENSOR SIGNAL PROCES
[5]  
Bin S, 2020, P 2020 INT C COMPUTI, P107
[6]   CMFA_Net: A cross-modal feature aggregation network for infrared-visible image fusion [J].
Ding, Zhaisheng ;
Li, Haiyan ;
Zhou, Dongming ;
Li, Hongsong ;
Liu, Yanyu ;
Hou, Ruichao .
INFRARED PHYSICS & TECHNOLOGY, 2021, 118
[7]   Face completion with Hybrid Dilated Convolution [J].
Fang, Yuchun ;
Li, Yifan ;
Tu, Xiaokang ;
Tan, Taifeng ;
Wang, Xin .
SIGNAL PROCESSING-IMAGE COMMUNICATION, 2020, 80
[8]   Colour Mapping: A Review of Recent Methods, Extensions and Applications [J].
Faridul, H. Sheikh ;
Pouli, T. ;
Chamaret, C. ;
Stauder, J. ;
Reinhard, E. ;
Kuzovkin, D. ;
Tremeau, A. .
COMPUTER GRAPHICS FORUM, 2016, 35 (01) :59-88
[9]   Directed color transfer for low-light image enhancement [J].
Florea, Laura ;
Florea, Corneliu .
DIGITAL SIGNAL PROCESSING, 2019, 93 :1-12
[10]   Efficient Heuristic Methods for Multimodal Fusion and Concept Fusion in Video Concept Detection [J].
Geng, Jie ;
Miao, Zhenjiang ;
Zhang, Xiao-Ping .
IEEE TRANSACTIONS ON MULTIMEDIA, 2015, 17 (04) :498-511