Image super-resolution reconstruction using Swin Transformer with efficient channel attention networks

被引:4
|
作者
Sun, Zhenxi [1 ,2 ]
Zhang, Jin [1 ,2 ,3 ]
Chen, Ziyi [1 ,2 ]
Hong, Lu [1 ,2 ]
Zhang, Rui [1 ,2 ]
Li, Weishi [1 ,2 ,3 ]
Xia, Haojie [1 ,2 ,3 ]
机构
[1] Hefei Univ Technol, Sch Instrument Sci & Optoelect Engn, Hefei 230009, Peoples R China
[2] Anhui Prov Key Lab Measuring Theory & Precis Instr, Hefei 230009, Peoples R China
[3] Minist Educ, Engn Res Ctr Safety Crit Ind Measurement & Control, Hefei 230009, Peoples R China
基金
国家重点研发计划;
关键词
Image super-resolution; Swin Transformer; Efficient channel attention; Multi-attention fusion;
D O I
10.1016/j.engappai.2024.108859
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Image super-resolution reconstruction (SR) is an important ill-posed problem in low-level vision, which aims to reconstruct high-resolution images from low-resolution images. Although current state-of-the-art methods exhibit impressive performance, their recovery of image detail information and edge information is still unsatisfactory. To address this problem, this paper proposes a shifted window Transformer (Swin Transformer) with an efficient channel attention network (S-ECAN), which combines the attention based on convolutional neural networks and the self-attention of the Swin Transformer to combine the advantages of both and focuses on learning high-frequency features of images. In addition, to solve the problem of Convolutional Neural Network (CNN) based channel attention consumes a large number of parameters to achieve good performance, this paper proposes the Efficient Channel Attention Block (ECAB), which only involves a handful of parameters while bringing clear performance gain. Extensive experimental validation shows that the proposed model can recover more high-frequency details and texture information. The model is validated on Set5, Set14, B100, Urban100, and Manga109 datasets, where it outperforms the state-of-the-art methods by 0.03-0.13 dB, 0.04- 0.09 dB, 0.01-0.06 dB, 0.13-0.20 dB, and 0.06-0.17 dB respectively in terms of objective metrics. Ultimately, the substantial performance gains and enhanced visual results over prior arts validate the effectiveness and competitiveness of our proposed approach, which achieves an improved performance-complexity trade-off.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Efficient image super-resolution integration
    Ke Xu
    Xin Wang
    Xin Yang
    Shengfeng He
    Qiang Zhang
    Baocai Yin
    Xiaopeng Wei
    Rynson W. H. Lau
    The Visual Computer, 2018, 34 : 1065 - 1076
  • [42] Lightweight Super-Resolution Image-Reconstruction Model with Adaptive Residual Attention
    Jiang Ming
    Xiao Qingsheng
    Yi Jianbing
    Cao Feng
    LASER & OPTOELECTRONICS PROGRESS, 2022, 59 (16)
  • [43] Image Super-Resolution via Efficient Transformer Embedding Frequency Decomposition With Restart
    Zuo, Yifan
    Yao, Wenhao
    Hu, Yuqi
    Fang, Yuming
    Liu, Wei
    Peng, Yuxin
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 4670 - 4685
  • [44] MadFormer: multi-attention-driven image super-resolution method based on Transformer
    Beibei Liu
    Jing Sun
    Bing Zhu
    Ting Li
    Fuming Sun
    Multimedia Systems, 2024, 30
  • [45] Super-resolution image reconstruction method using homotopy regularization
    Wang, Liping
    Zhou, Shangbo
    Karim, Awudu
    MULTIMEDIA TOOLS AND APPLICATIONS, 2016, 75 (23) : 15993 - 16016
  • [46] Super-resolution image reconstruction method using homotopy regularization
    Liping Wang
    Shangbo Zhou
    Awudu Karim
    Multimedia Tools and Applications, 2016, 75 : 15993 - 16016
  • [47] MadFormer: multi-attention-driven image super-resolution method based on Transformer
    Liu, Beibei
    Sun, Jing
    Zhu, Bing
    Li, Ting
    Sun, Fuming
    MULTIMEDIA SYSTEMS, 2024, 30 (02)
  • [48] High-frequency channel attention and contrastive learning for image super-resolution
    Yan, Tianyu
    Yin, Hujun
    VISUAL COMPUTER, 2024, 40 (12) : 8839 - 8851
  • [49] Super-resolution image reconstruction for mobile devices
    Chu, Chung-Hua
    MULTIMEDIA SYSTEMS, 2013, 19 (04) : 315 - 337
  • [50] Enhancing Image Super-Resolution with Dual Compression Transformer
    Yu, Jiaxing
    Chen, Zheng
    Wang, Jingkai
    Kong, Linghe
    Yan, Jiajie
    Gu, Wei
    VISUAL COMPUTER, 2025, 41 (07) : 4879 - 4892