GTMFuse: Group-attention transformer-driven multiscale dense feature-enhanced network for infrared and visible image fusion

被引:16
|
作者
Mei, Liye [1 ,2 ]
Hu, Xinglong [1 ]
Ye, Zhaoyi [1 ]
Tang, Linfeng [3 ]
Wang, Ying [4 ]
Li, Di [1 ]
Liu, Yan [4 ]
Hao, Xin [5 ]
Lei, Cheng [2 ]
Xu, Chuan [1 ]
Yang, Wei [4 ,6 ]
机构
[1] Hubei Univ Technol, Sch Comp Sci, Wuhan 430068, Peoples R China
[2] Wuhan Univ, Inst Technol Sci, Wuhan 430072, Peoples R China
[3] Wuhan Univ, Elect Informat Sch, Wuhan 430072, Peoples R China
[4] Wuchang Shouyi Univ, Sch Informat Sci & Engn, Wuhan 430064, Peoples R China
[5] Antgroup, Hangzhou 310020, Peoples R China
[6] Wuhan Univ, State Key Lab Informat Engn Surveying Mapping & Re, Wuhan 430072, Peoples R China
关键词
Infrared and visible image fusion; Deep learning; Group; -attention; Multiscale feature;
D O I
10.1016/j.knosys.2024.111658
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Infrared and visible images captured by different devices can be seamlessly integrated into a single composite image through the application of image fusion techniques. However, many existing convolutional neural network-based methods for infrared and visible image fusion have exhibited limited capability for effectively amalgamating information from the source images. Consequently, we propose a group-attention transformer into the multiscale feature enhanced network for infrared and visible image fusion, which we abbreviate as GTMFuse. Specifically, GTMFuse employs multiscale dual-channel encoders to independently process the source image and extract multiscale features. Among the encoders, the group-attention transformer module is utilized to facilitate more comprehensive long-range feature dependency modeling at each scale. This innovative module seamlessly combines a fixed-direction stripe attention mechanism with channel attention and window attention, enabling comprehensive global long-range information capture and interaction with feature information across the source images. The multiscale features obtained from the group-attention transformer module are integrated into the fused image through a meticulously designed dense fusion block. Furthermore, this study introduces a novel dataset named HBUT-IV, encompassing surveillance images captured from multiple viewpoints. The HBUT-IV dataset serves as a valuable benchmark for assessing the efficacy of fusion methods. Extensive experiments are conducted on four datasets employing nine comparative methods, revealing the superior performance of the GTMFuse approach. The implementation code is accessible at https://github.com/XingLongH/GTMFuse.
引用
收藏
页数:15
相关论文
共 50 条
  • [41] Infrared and Visible Image Fusion via Attention-Based Adaptive Feature Fusion
    Wang, Lei
    Hu, Ziming
    Kong, Quan
    Qi, Qian
    Liao, Qing
    ENTROPY, 2023, 25 (03)
  • [42] RITFusion: Reinforced Interactive Transformer Network for Infrared and Visible Image Fusion
    Li, Xiaoling
    Li, Yanfeng
    Chen, Houjin
    Peng, Yahui
    Chen, Luyifu
    Wang, Minjun
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73
  • [43] RXDNFuse: A aggregated residual dense network for infrared and visible image fusion
    Long, Yongzhi
    Jia, Haitao
    Zhong, Yida
    Jiang, Yadong
    Jia, Yuming
    INFORMATION FUSION, 2021, 69 : 128 - 141
  • [44] MTDFusion: A Multilayer Triple Dense Network for Infrared and Visible Image Fusion
    Karim, Shahid
    Tong, Geng
    Li, Jinyang
    Yu, Xiaochang
    Hao, Jia
    Qadir, Akeel
    Yu, Yiting
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 17
  • [45] Infrared and visible image fusion with deep wavelet-dense network
    Chen, Yanling
    Cheng, Lianglun
    Wu, Heng
    Chen, Ziyang
    LI, Feng
    OPTICA APPLICATA, 2023, 53 (01) : 49 - 64
  • [46] Global attention network with multiscale feature fusion for infrared small target detection
    Zhang, Fan
    Lin, Shunlong
    Xiao, Xiaoyang
    Wang, Yun
    Zhao, Yuqian
    OPTICS AND LASER TECHNOLOGY, 2024, 168
  • [47] Infrared and Visible Image Fusion via Multiscale Receptive Field Amplification Fusion Network
    Ji, Chuanming
    Zhou, Wujie
    Lei, Jingsheng
    Ye, Lv
    IEEE SIGNAL PROCESSING LETTERS, 2023, 30 : 493 - 497
  • [48] Self-Attention Progressive Network for Infrared and Visible Image Fusion
    Li, Shuying
    Han, Muyi
    Qin, Yuemei
    Li, Qiang
    REMOTE SENSING, 2024, 16 (18)
  • [49] Infrared and visible image fusion based on dilated residual attention network
    Mustafa, Hafiz Tayyab
    Yang, Jie
    Mustafa, Hamza
    Zareapoor, Masoumeh
    OPTIK, 2020, 224 (224):
  • [50] Attention based dual UNET network for infrared and visible image fusion
    Wang, Xuejiao
    Hua, Zhen
    Li, Jinjiang
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (25) : 66959 - 66980