共 50 条
ESFuse: Weak Edge Structure Perception Network for Infrared and Visible Image Fusion
被引:1
|作者:
Liu, Wuyang
[1
,2
]
Tan, Haishu
[3
,4
]
Cheng, Xiaoqi
[1
,2
]
Li, Xiaosong
[2
,3
,4
]
机构:
[1] Foshan Univ, Sch Mechatron Engn & Automat, Foshan 528000, Peoples R China
[2] Foshan Univ, Guangdong Prov Key Lab Ind Intelligent Inspect Tec, Foshan 528000, Peoples R China
[3] Foshan Univ, Sch Phys & Optoelect Engn, Foshan 528225, Peoples R China
[4] Foshan Univ, Guangdong HongKong Macao Joint Lab Intelligent Mic, Foshan 528225, Guangdong, Peoples R China
来源:
基金:
中国国家自然科学基金;
关键词:
infrared and visible image fusion;
weak edge structure perception;
multiscale feature;
INFORMATION MEASURE;
PERFORMANCE;
ALGORITHM;
MODEL;
D O I:
10.3390/electronics13204115
中图分类号:
TP [自动化技术、计算机技术];
学科分类号:
0812 ;
摘要:
Infrared and visible image fusion (IVIF) fully integrates the complementary features of different modal images, and the fused image provides a more comprehensive and objective interpretation of the scene compared to each source image, thus attracting extensive attention in the field of computer vision in recent years. However, current fusion methods usually center their attention on the extraction of prominent features, falling short of adequately safeguarding subtle and diminutive structures. To address this problem, we propose an end-to-end unsupervised IVIF method (ESFuse), which effectively enhances fine edges and small structures. In particular, we introduce a two-branch head interpreter to extract features from source images of different modalities. Subsequently, these features are fed into the edge refinement module with the detail injection module (DIM) to obtain the edge detection results of the source image, improving the network's ability to capture and retain complex details as well as global information. Finally, we implemented a multiscale feature reconstruction module to obtain the final fusion results by combining the output of the DIM with the output of the head interpreter. Extensive IVIF fusion experiments on existing publicly available datasets show that the proposed ESFuse outperforms the state-of-the-art(SOTA) methods in both subjective vision and objective evaluation, and our fusion results perform well in semantic segmentation, target detection, pose estimation and depth estimation tasks. The source code has been availabled.
引用
收藏
页数:21
相关论文