ReCoNet: Recurrent Correction Network for Fast and Efficient Multi-modality Image Fusion

被引:109
作者
Huang, Zhanbo [1 ]
Liu, Jinyuan [2 ]
Fan, Xin [1 ]
Liu, Risheng [1 ,3 ]
Zhong, Wei [1 ]
Luo, Zhongxuan [1 ]
机构
[1] Dalian Univ Technol, DUT RU Int Sch Informat Sci & Engn, Dalian, Peoples R China
[2] Dalian Univ Technol, Sch Software Technol, Dalian, Peoples R China
[3] Peng Cheng Lab, Shenzhen, Peoples R China
来源
COMPUTER VISION - ECCV 2022, PT XVIII | 2022年 / 13678卷
关键词
Deep learning; Multi-modality image fusion; VISION;
D O I
10.1007/978-3-031-19797-0_31
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advances in deep networks have gained great attention in infrared and visible image fusion (IVIF). Nevertheless, most existing methods are incapable of dealing with slight misalignment on source images and suffer from high computational and spatial expenses. This paper tackles these two critical issues rarely touched in the community by developing a recurrent correction network for robust and efficient fusion, namely ReCoNet. Concretely, we design a deformation module to explicitly compensate geometrical distortions and an attention mechanism to mitigate ghosting-like artifacts, respectively. Meanwhile, the network consists of a parallel dilated convolutional layer and runs in a recurrent fashion, significantly reducing both spatial and computational complexities. ReCoNet can effectively and efficiently alleviates both structural distortions and textural artifacts brought by slight misalignment. Extensive experiments on two public datasets demonstrate the superior accuracy and efficacy of our ReCoNet against the state-of-the-art IVIF methods. Consequently, we obtain a 16% relative improvement of CC on datasets with misalignment and boost the efficiency by 86%. The source code is available at https://github.com/dlut-dimt/reconet.
引用
收藏
页码:539 / 555
页数:17
相关论文
共 47 条
[1]  
Bochkovskiy A, 2020, Arxiv, DOI [arXiv:2004.10934, DOI 10.48550/ARXIV.2004.10934, 10.48550/arXiv.2004.10934]
[2]   Learning a Neural Solver for Multiple Object Tracking [J].
Braso, Guillem ;
Leal-Taixe, Laura .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :6246-6256
[3]   The Cityscapes Dataset for Semantic Urban Scene Understanding [J].
Cordts, Marius ;
Omran, Mohamed ;
Ramos, Sebastian ;
Rehfeld, Timo ;
Enzweiler, Markus ;
Benenson, Rodrigo ;
Franke, Uwe ;
Roth, Stefan ;
Schiele, Bernt .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :3213-3223
[4]   Fusing Infrared and Visible Images of Different Resolutions via Total Variation Model [J].
Du, Qinglei ;
Xu, Han ;
Ma, Yong ;
Huang, Jun ;
Fan, Fan .
SENSORS, 2018, 18 (11)
[5]   Dual Attention Network for Scene Segmentation [J].
Fu, Jun ;
Liu, Jing ;
Tian, Haijie ;
Li, Yong ;
Bao, Yongjun ;
Fang, Zhiwei ;
Lu, Hanqing .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :3141-3149
[6]   Object Classification Using CNN-Based Fusion of Vision and LIDAR in Autonomous Vehicle Environment [J].
Gao, Hongbo ;
Cheng, Bo ;
Wang, Jianqiang ;
Li, Keqiang ;
Zhao, Jianhui ;
Li, Deyi .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2018, 14 (09) :4224-4231
[7]   Vision meets robotics: The KITTI dataset [J].
Geiger, A. ;
Lenz, P. ;
Stiller, C. ;
Urtasun, R. .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2013, 32 (11) :1231-1237
[8]   Unsupervised Monocular Depth Estimation with Left-Right Consistency [J].
Godard, Clement ;
Mac Aodha, Oisin ;
Brostow, Gabriel J. .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6602-6611
[9]  
He KM, 2017, IEEE I CONF COMP VIS, P2980, DOI [10.1109/TPAMI.2018.2844175, 10.1109/ICCV.2017.322]
[10]   Target Oriented Perceptual Adversarial Fusion Network for Underwater Image Enhancement [J].
Jiang, Zhiying ;
Li, Zhuoxiao ;
Yang, Shuzhou ;
Fan, Xin ;
Liu, Risheng .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (10) :6584-6598