Efficient Dual Attention Transformer for Image Super-Resolution

被引:0
|
作者
Park, Soobin [1 ]
Jeong, Yuna [1 ]
Choi, Yong Suk [1 ]
机构
[1] Hanyang Univ, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
Image super-resolution; Low-level vision; Vision transformer; Self-attention; Computer vision;
D O I
10.1145/3605098.3635991
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Research based on computationally efficient local-window self-attention has been actively advancing in the field of image super-resolution (SR), leading to significant performance improvements. However, in most recent studies, local-window self-attention focuses only on spatial dimension, without sufficient consideration of the channel dimension. Additionally, extracting global information while maintaining the efficiency of local-window self-attention, still remains a challenging task in image SR. To resolve these problems, we propose a novel efficient dual attention transformer (EDAT). Our EDAT presents a dual attention block (DAB) that empowers the exploration of interdependencies not just among features residing at diverse spatial locations but also among distinct channels. Moreover, we propose a global attention block (GAB) to achieve efficient global feature extraction by reducing the spatial size of the keys and values. Our extensive experiments demonstrate that our DAB and GAB complement each other, exhibiting a synergistic effect. Furthermore, based on the two attention blocks, DAB and GAB, our EDAT achieves state-of-the-art results on five benchmark datasets.
引用
收藏
页码:963 / 970
页数:8
相关论文
共 50 条
  • [31] Dual Attention with the Self-Attention Alignment for Efficient Video Super-resolution
    Chu, Yuezhong
    Qiao, Yunan
    Liu, Heng
    Han, Jungong
    COGNITIVE COMPUTATION, 2022, 14 (03) : 1140 - 1151
  • [32] HiT-SR: Hierarchical Transformer for Efficient Image Super-Resolution
    Zhang, Xiang
    Zhang, Yulun
    Yu, Fisher
    COMPUTER VISION - ECCV 2024, PT XL, 2025, 15098 : 483 - 500
  • [33] Transformer-Based Selective Super-resolution for Efficient Image Refinement
    Zhang, Tianyi
    Kasichainula, Kishore
    Zhuo, Yaoxin
    Li, Baoxin
    Seo, Jae-Sun
    Cao, Yu
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 7, 2024, : 7305 - 7313
  • [34] Asymmetric content-aided Transformer for efficient image super-resolution
    Wang, Qian
    Mao, Yanyu
    Guo, Ruilong
    Tang, Yao
    Wei, Jing
    Quan, Bo
    KNOWLEDGE-BASED SYSTEMS, 2025, 315
  • [35] AdaFormer: Efficient Transformer with Adaptive Token Sparsification for Image Super-resolution
    Luo, Xiaotong
    Ai, Zekun
    Liang, Qiuyuan
    Liu, Ding
    Xie, Yuan
    Qu, Yanyun
    Fu, Yun
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 5, 2024, : 4009 - 4016
  • [36] EdgeFormer: Edge-Aware Efficient Transformer for Image Super-Resolution
    Luo, Xiaotong
    Ai, Zekun
    Liang, Qiuyuan
    Xie, Yuan
    Shi, Zhongchao
    Fan, Jianping
    Qu, Yanyun
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73
  • [37] Super-Resolution Generative Adversarial Network Based on the Dual Dimension Attention Mechanism for Biometric Image Super-Resolution
    Huang, Chi-En
    Li, Yung-Hui
    Aslam, Muhammad Saqlain
    Chang, Ching-Chun
    SENSORS, 2021, 21 (23)
  • [38] EHAT:Enhanced Hybrid Attention Transformer for Remote Sensing Image Super-Resolution
    Wang, Jian
    Xie, Zexin
    Du, Yanlin
    Song, Wei
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT VIII, 2025, 15038 : 225 - 237
  • [39] Multi-attention fusion transformer for single-image super-resolution
    Li, Guanxing
    Cui, Zhaotong
    Li, Meng
    Han, Yu
    Li, Tianping
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [40] Joint image deblurring and super-resolution with attention dual supervised network
    Zhang, Dongyang
    Liang, Zhenwen
    Shao, Jie
    NEUROCOMPUTING, 2020, 412 (412) : 187 - 196