MadFormer: multi-attention-driven image super-resolution method based on Transformer

被引:4
|
作者
Liu, Beibei [1 ]
Sun, Jing [1 ]
Zhu, Bing [2 ]
Li, Ting [1 ]
Sun, Fuming [1 ]
机构
[1] Dalian Minzu Univ, Sch Informat & Commun Engn, Liaohe West Rd, Dalian 116600, Liaoning, Peoples R China
[2] Harbin Inst Technol, Sch Elect & Informat Engn, Xidazhi St, Harbin 150006, Heilongjiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Image super-resolution; Transformer; Multi-attention-driven; Dynamic fusion;
D O I
10.1007/s00530-024-01276-1
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
While the Transformer-based method has demonstrated exceptional performance in low-level visual processing tasks, it has a strong modeling ability only locally, thereby neglecting the importance of spatial feature information and high-frequency details within the channel for super-resolution. To enhance feature information and improve the visual experience, we propose a multi-attention-driven image super-resolution method based on a Transformer network, called MadFormer. Initially, the low-resolution image undergoes an initial convolution operation to extract shallow features while being fed into a residual multi-attention block incorporating channel attention, spatial attention, and self-attention mechanisms. By employing multi-head self-attention, the proposed method aims to capture global-local feature information; channel attention and spatial attention are utilized to effectively capture high-frequency features in both the channel and spatial domains. Subsequently, deep feature information is inputted into a dynamic fusion block that dynamically fuses multi-attention extracted features, facilitating the aggregation of cross-window information. Ultimately, the shallow and deep feature information is fused via convolution operations, yielding high-resolution images through high-quality reconstruction. Comprehensive quantitative and qualitative comparisons with other advanced algorithms demonstrate the substantial advantages of the proposed approach in terms of peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) for image super-resolution.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] Image super-resolution network based on a multi-branch attention mechanism
    Yang, Xin
    Guo, Yingqing
    Li, Zhiqiang
    Zhou, Dake
    SIGNAL IMAGE AND VIDEO PROCESSING, 2021, 15 (07) : 1397 - 1405
  • [22] Attention-based Multi-Reference Learning for Image Super-Resolution
    Pesavento, Marco
    Volino, Marco
    Hilton, Adrian
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 14677 - 14686
  • [23] Image super-resolution network based on multi-scale adaptive attention
    Zhou Y.
    Pei S.
    Chen H.
    Xu S.
    Guangxue Jingmi Gongcheng/Optics and Precision Engineering, 2024, 32 (06): : 843 - 856
  • [24] Multi-scale attention network for image super-resolution
    Wang, Li
    Shen, Jie
    Tang, E.
    Zheng, Shengnan
    Xu, Lizhong
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2021, 80
  • [25] Degradation-Aware Self-Attention Based Transformer for Blind Image Super-Resolution
    Liu, Qingguo
    Gao, Pan
    Han, Kang
    Liu, Ningzhong
    Xiang, Wei
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 7516 - 7528
  • [26] MSWAGAN: Multispectral Remote Sensing Image Super-Resolution Based on Multiscale Window Attention Transformer
    Wang, Chunyang
    Zhang, Xian
    Yang, Wei
    Wang, Gaige
    Li, Xingwang
    Wang, Jianlong
    Lu, Bibo
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 15
  • [27] A Lightweight Hyperspectral Image Super-Resolution Method Based on Multiple Attention Mechanisms
    Bu, Lijing
    Dai, Dong
    Zhang, Zhengpeng
    Xie, Xinyu
    Deng, Mingjun
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, ICIC 2023, PT II, 2023, 14087 : 639 - 651
  • [28] MAFT: An Image Super-Resolution Method Based on Mixed Attention and Feature Transfer
    Liu, Xin
    Li, Jing
    Cui, Yuanning
    Zhu, Wei
    Qian, Luhong
    WEB AND BIG DATA, PT II, APWEB-WAIM 2022, 2023, 13422 : 511 - 519
  • [29] EHAT:Enhanced Hybrid Attention Transformer for Remote Sensing Image Super-Resolution
    Wang, Jian
    Xie, Zexin
    Du, Yanlin
    Song, Wei
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT VIII, 2025, 15038 : 225 - 237
  • [30] Dual Self-Attention Swin Transformer for Hyperspectral Image Super-Resolution
    Long, Yaqian
    Wang, Xun
    Xu, Meng
    Zhang, Shuyu
    Jiang, Shuguo
    Jia, Sen
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61