Image super-resolution reconstruction using Swin Transformer with efficient channel attention networks

被引:4
|
作者
Sun, Zhenxi [1 ,2 ]
Zhang, Jin [1 ,2 ,3 ]
Chen, Ziyi [1 ,2 ]
Hong, Lu [1 ,2 ]
Zhang, Rui [1 ,2 ]
Li, Weishi [1 ,2 ,3 ]
Xia, Haojie [1 ,2 ,3 ]
机构
[1] Hefei Univ Technol, Sch Instrument Sci & Optoelect Engn, Hefei 230009, Peoples R China
[2] Anhui Prov Key Lab Measuring Theory & Precis Instr, Hefei 230009, Peoples R China
[3] Minist Educ, Engn Res Ctr Safety Crit Ind Measurement & Control, Hefei 230009, Peoples R China
基金
国家重点研发计划;
关键词
Image super-resolution; Swin Transformer; Efficient channel attention; Multi-attention fusion;
D O I
10.1016/j.engappai.2024.108859
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Image super-resolution reconstruction (SR) is an important ill-posed problem in low-level vision, which aims to reconstruct high-resolution images from low-resolution images. Although current state-of-the-art methods exhibit impressive performance, their recovery of image detail information and edge information is still unsatisfactory. To address this problem, this paper proposes a shifted window Transformer (Swin Transformer) with an efficient channel attention network (S-ECAN), which combines the attention based on convolutional neural networks and the self-attention of the Swin Transformer to combine the advantages of both and focuses on learning high-frequency features of images. In addition, to solve the problem of Convolutional Neural Network (CNN) based channel attention consumes a large number of parameters to achieve good performance, this paper proposes the Efficient Channel Attention Block (ECAB), which only involves a handful of parameters while bringing clear performance gain. Extensive experimental validation shows that the proposed model can recover more high-frequency details and texture information. The model is validated on Set5, Set14, B100, Urban100, and Manga109 datasets, where it outperforms the state-of-the-art methods by 0.03-0.13 dB, 0.04- 0.09 dB, 0.01-0.06 dB, 0.13-0.20 dB, and 0.06-0.17 dB respectively in terms of objective metrics. Ultimately, the substantial performance gains and enhanced visual results over prior arts validate the effectiveness and competitiveness of our proposed approach, which achieves an improved performance-complexity trade-off.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Efficient Swin Transformer for Remote Sensing Image Super-Resolution
    Kang, Xudong
    Duan, Puhong
    Li, Jier
    Li, Shutao
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 6367 - 6379
  • [2] Efficient Dual Attention Transformer for Image Super-Resolution
    Park, Soobin
    Jeong, Yuna
    Choi, Yong Suk
    39TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, SAC 2024, 2024, : 963 - 970
  • [3] SwinT-SRNet: Swin transformer with image super-resolution reconstruction network for pollen images classification
    Zu, Baokai
    Cao, Tong
    Li, Yafang
    Li, Jianqiang
    Ju, Fujiao
    Wang, Hongyuan
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133
  • [4] Asymmetric convolution Swin transformer for medical image super-resolution
    Lu, Weijia
    Jiang, Jiehui
    Tian, Hao
    Gu, Jun
    Lu, Yuhong
    Yang, Wanli
    Gong, Ming
    Han, Tianyi
    Jiang, Xiaojuan
    Zhang, Tingting
    ALEXANDRIA ENGINEERING JOURNAL, 2023, 85 : 177 - 184
  • [5] Image super-resolution using dilated neighborhood attention transformer
    Chen, Li
    Zuo, Jinnian
    Du, Kai
    Zou, Jinsong
    Yin, Shaoyun
    Wang, Jinyu
    JOURNAL OF ELECTRONIC IMAGING, 2024, 33 (02)
  • [6] STGAN: Swin Transformer-Based GAN to Achieve Remote Sensing Image Super-Resolution Reconstruction
    Huo, Wei
    Zhang, Xiaodan
    You, Shaojie
    Zhang, Yongkun
    Zhang, Qiyuan
    Hu, Naihao
    APPLIED SCIENCES-BASEL, 2025, 15 (01):
  • [7] Single Image Super-Resolution Using Feedback Attention Networks
    Zhang, Juntao
    Dong, Hongbin
    Huang, Ruolin
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 2808 - 2814
  • [8] Image Super-Resolution Reconstruction Based on Dual-Branch Channel Attention
    Shi, Jinyu
    Si, Zhanjun
    Zhang, Yingxue
    Yang, Xinbin
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT VII, ICIC 2024, 2024, 14868 : 291 - 299
  • [9] HCT: image super-resolution restoration using hierarchical convolution transformer networks
    Guo, Ying
    Tian, Chang
    Wang, Han
    Liu, Jie
    Di, Chong
    Ning, Keqing
    PATTERN ANALYSIS AND APPLICATIONS, 2025, 28 (02)
  • [10] A SWIN TRANSFORMER- BASED FUSION APPROACH FOR HYPERSPECTRAL IMAGE SUPER-RESOLUTION
    Yang, Yuchao
    Wang, Yulei
    Zhao, Enyu
    Song, Meiping
    Zhang, Qiang
    IGARSS 2023 - 2023 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, 2023, : 7372 - 7375