Image super-resolution reconstruction using Swin Transformer with efficient channel attention networks

被引:4
|
作者
Sun, Zhenxi [1 ,2 ]
Zhang, Jin [1 ,2 ,3 ]
Chen, Ziyi [1 ,2 ]
Hong, Lu [1 ,2 ]
Zhang, Rui [1 ,2 ]
Li, Weishi [1 ,2 ,3 ]
Xia, Haojie [1 ,2 ,3 ]
机构
[1] Hefei Univ Technol, Sch Instrument Sci & Optoelect Engn, Hefei 230009, Peoples R China
[2] Anhui Prov Key Lab Measuring Theory & Precis Instr, Hefei 230009, Peoples R China
[3] Minist Educ, Engn Res Ctr Safety Crit Ind Measurement & Control, Hefei 230009, Peoples R China
基金
国家重点研发计划;
关键词
Image super-resolution; Swin Transformer; Efficient channel attention; Multi-attention fusion;
D O I
10.1016/j.engappai.2024.108859
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Image super-resolution reconstruction (SR) is an important ill-posed problem in low-level vision, which aims to reconstruct high-resolution images from low-resolution images. Although current state-of-the-art methods exhibit impressive performance, their recovery of image detail information and edge information is still unsatisfactory. To address this problem, this paper proposes a shifted window Transformer (Swin Transformer) with an efficient channel attention network (S-ECAN), which combines the attention based on convolutional neural networks and the self-attention of the Swin Transformer to combine the advantages of both and focuses on learning high-frequency features of images. In addition, to solve the problem of Convolutional Neural Network (CNN) based channel attention consumes a large number of parameters to achieve good performance, this paper proposes the Efficient Channel Attention Block (ECAB), which only involves a handful of parameters while bringing clear performance gain. Extensive experimental validation shows that the proposed model can recover more high-frequency details and texture information. The model is validated on Set5, Set14, B100, Urban100, and Manga109 datasets, where it outperforms the state-of-the-art methods by 0.03-0.13 dB, 0.04- 0.09 dB, 0.01-0.06 dB, 0.13-0.20 dB, and 0.06-0.17 dB respectively in terms of objective metrics. Ultimately, the substantial performance gains and enhanced visual results over prior arts validate the effectiveness and competitiveness of our proposed approach, which achieves an improved performance-complexity trade-off.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Medical-Image Super-Resolution Reconstruction Method Based on Residual Channel Attention Network
    Liu Kewen
    Ma Yuan
    Xiong Hongxia
    Yan Zejun
    Zhou Zhijun
    Liu Chaoyang
    Fang Panpan
    Li Xiaojun
    Chen Yalei
    LASER & OPTOELECTRONICS PROGRESS, 2020, 57 (02)
  • [22] A Lightweight Single-Image Super-Resolution Method Based on the Parallel Connection of Convolution and Swin Transformer Blocks
    Jing, Tengyun
    Liu, Cuiyin
    Chen, Yuanshuai
    APPLIED SCIENCES-BASEL, 2025, 15 (04):
  • [23] Efficient residual attention network for single image super-resolution
    Fangwei Hao
    Taiping Zhang
    Linchang Zhao
    Yuanyan Tang
    Applied Intelligence, 2022, 52 : 652 - 661
  • [24] Efficient residual attention network for single image super-resolution
    Hao, Fangwei
    Zhang, Taiping
    Zhao, Linchang
    Tang, Yuanyan
    APPLIED INTELLIGENCE, 2022, 52 (01) : 652 - 661
  • [25] Spatial and channel enhanced self-attention network for efficient single image super-resolution
    Song, Xiaogang
    Tan, Yuping
    Pang, Xinchao
    Zhang, Lei
    Lu, Xiaofeng
    Hei, Xinhong
    NEUROCOMPUTING, 2025, 620
  • [26] Stacked U-shape networks with channel-wise attention for image super-resolution
    Zhu, Leilei
    Zhan, Shu
    Zhang, Haiyan
    NEUROCOMPUTING, 2019, 345 : 58 - 66
  • [27] RSTSRN: Recursive Swin Transformer Super-Resolution Network for Mars Images
    Wu, Fanlu
    Jiang, Xiaonan
    Fu, Tianjiao
    Fu, Yao
    Xu, Dongdong
    Zhao, Chunlei
    APPLIED SCIENCES-BASEL, 2024, 14 (20):
  • [28] Image Super-Resolution Reconstruction Based on the Lightweight Hybrid Attention Network
    Chu, Yuezhong
    Wang, Kang
    Zhang, Xuefeng
    Heng, Liu
    ADVANCES IN MULTIMEDIA, 2024, 2024
  • [29] Multi-attention fusion transformer for single-image super-resolution
    Li, Guanxing
    Cui, Zhaotong
    Li, Meng
    Han, Yu
    Li, Tianping
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [30] Channel Graph Convolutional Networks for Animation Image Super-Resolution
    Wang, Fuchun
    Wang, Kesheng
    Song, Lei
    IEEE ACCESS, 2024, 12 : 197577 - 197588