Image super-resolution reconstruction using Swin Transformer with efficient channel attention networks

被引:4
|
作者
Sun, Zhenxi [1 ,2 ]
Zhang, Jin [1 ,2 ,3 ]
Chen, Ziyi [1 ,2 ]
Hong, Lu [1 ,2 ]
Zhang, Rui [1 ,2 ]
Li, Weishi [1 ,2 ,3 ]
Xia, Haojie [1 ,2 ,3 ]
机构
[1] Hefei Univ Technol, Sch Instrument Sci & Optoelect Engn, Hefei 230009, Peoples R China
[2] Anhui Prov Key Lab Measuring Theory & Precis Instr, Hefei 230009, Peoples R China
[3] Minist Educ, Engn Res Ctr Safety Crit Ind Measurement & Control, Hefei 230009, Peoples R China
基金
国家重点研发计划;
关键词
Image super-resolution; Swin Transformer; Efficient channel attention; Multi-attention fusion;
D O I
10.1016/j.engappai.2024.108859
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Image super-resolution reconstruction (SR) is an important ill-posed problem in low-level vision, which aims to reconstruct high-resolution images from low-resolution images. Although current state-of-the-art methods exhibit impressive performance, their recovery of image detail information and edge information is still unsatisfactory. To address this problem, this paper proposes a shifted window Transformer (Swin Transformer) with an efficient channel attention network (S-ECAN), which combines the attention based on convolutional neural networks and the self-attention of the Swin Transformer to combine the advantages of both and focuses on learning high-frequency features of images. In addition, to solve the problem of Convolutional Neural Network (CNN) based channel attention consumes a large number of parameters to achieve good performance, this paper proposes the Efficient Channel Attention Block (ECAB), which only involves a handful of parameters while bringing clear performance gain. Extensive experimental validation shows that the proposed model can recover more high-frequency details and texture information. The model is validated on Set5, Set14, B100, Urban100, and Manga109 datasets, where it outperforms the state-of-the-art methods by 0.03-0.13 dB, 0.04- 0.09 dB, 0.01-0.06 dB, 0.13-0.20 dB, and 0.06-0.17 dB respectively in terms of objective metrics. Ultimately, the substantial performance gains and enhanced visual results over prior arts validate the effectiveness and competitiveness of our proposed approach, which achieves an improved performance-complexity trade-off.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] Residual Swin Transformer Channel Attention Network for Image Demosaicing
    Xing, Wenzhu
    Egiazarian, Karen
    2022 10TH EUROPEAN WORKSHOP ON VISUAL INFORMATION PROCESSING (EUVIP), 2022,
  • [32] Channel Attention Network for Wireless Capsule Endoscopy Image Super-Resolution
    Sarvaiya, Anjali
    Vaghela, Hiren
    Upla, Kishor
    Raja, Kiran
    Pedersen, Marius
    COMPUTER VISION AND IMAGE PROCESSING, CVIP 2023, PT II, 2024, 2010 : 432 - 444
  • [33] SWCGAN: Generative Adversarial Network Combining Swin Transformer and CNN for Remote Sensing Image Super-Resolution
    Tu, Jingzhi
    Mei, Gang
    Ma, Zhengjing
    Piccialli, Francesco
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2022, 15 : 5662 - 5673
  • [34] ESTUGAN: Enhanced Swin Transformer with U-Net Discriminator for Remote Sensing Image Super-Resolution
    Yu, Chunhe
    Hong, Lingyue
    Pan, Tianpeng
    Li, Yufeng
    Li, Tingting
    ELECTRONICS, 2023, 12 (20)
  • [35] Spatial relaxation transformer for image super-resolution
    Li, Yinghua
    Zhang, Ying
    Zeng, Hao
    He, Jinglu
    Guo, Jie
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2024, 36 (07)
  • [36] Residual attention network using multi-channel dense connections for image super-resolution
    Liu, Zhiwei
    Huang, Ji
    Zhu, Chengjia
    Peng, Xiaoyu
    Du, Xinyu
    APPLIED INTELLIGENCE, 2021, 51 (01) : 85 - 99
  • [37] Residual attention network using multi-channel dense connections for image super-resolution
    Zhiwei Liu
    Ji Huang
    Chengjia Zhu
    Xiaoyu Peng
    Xinyu Du
    Applied Intelligence, 2021, 51 : 85 - 99
  • [38] Efficient image super-resolution integration
    Xu, Ke
    Wang, Xin
    Yang, Xin
    He, Shengfeng
    Zhang, Qiang
    Yin, Baocai
    Wei, Xiaopeng
    Lau, Rynson W. H.
    VISUAL COMPUTER, 2018, 34 (6-8) : 1065 - 1076
  • [39] Efficient Blind Image Super-Resolution
    Vais, Olga
    Makarov, Ilya
    ADVANCES IN COMPUTATIONAL INTELLIGENCE, IWANN 2023, PT II, 2023, 14135 : 229 - 240
  • [40] Image super-resolution reconstruction under partial convolution and agent attention mechanism
    Chen, Yupeng
    Li, Haibo
    JOURNAL OF ELECTRONIC IMAGING, 2024, 33 (06)