Joint features-guided linear transformer and CNN for efficient image super-resolution

被引:2
|
作者
Wang, Bufan [1 ]
Zhang, Yongjun [1 ]
Long, Wei [1 ]
Cui, Zhongwei [2 ]
机构
[1] Guizhou Univ, Coll Comp Sci & Technol, State Key Lab Publ Big Data, Guiyang 550025, Guizhou, Peoples R China
[2] Guizhou Educ Univ, Sch Math & Big Data, Guiyang 550018, Peoples R China
关键词
Image super-resolution; Multi-level contextual information; Linear self-attention; Lightweight network; NETWORK;
D O I
10.1007/s13042-024-02277-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Integrating convolutional neural networks (CNNs) and transformers has notably improved lightweight single image super-resolution (SISR) tasks. However, existing methods lack the capability to exploit multi-level contextual information, and transformer computations inherently add quadratic complexity. To address these issues, we propose a Joint features-Guided Linear Transformer and CNN Network (JGLTN) for efficient SISR, which is constructed by cascading modules composed of CNN layers and linear transformer layers. Specifically, in the CNN layer, our approach employs an inter-scale feature integration module (IFIM) to extract critical latent information across scales. Then, in the linear transformer layer, we design a joint feature-guided linear attention (JGLA). It jointly considers adjacent and extended regional features, dynamically assigning weights to convolutional kernels for contextual feature selection. This process garners multi-level contextual information, which is used to guide linear attention for effective information interaction. Moreover, we redesign the method of computing feature similarity within the self-attention, reducing its computational complexity to linear. Extensive experiments shows that our proposal outperforms state-of-the-art models while balancing performance and computational costs.
引用
收藏
页码:5765 / 5780
页数:16
相关论文
共 50 条
  • [1] Image super-resolution method based on the interactive fusion of transformer and CNN features
    Wang, Jianxin
    Zou, Yongsong
    Alfarraj, Osama
    Sharma, Pradip Kumar
    Said, Wael
    Wang, Jin
    VISUAL COMPUTER, 2024, 40 (08) : 5827 - 5839
  • [2] Steformer: Efficient Stereo Image Super-Resolution With Transformer
    Lin, Jianxin
    Yin, Lianying
    Wang, Yijun
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 8396 - 8407
  • [3] An Efficient Latent Style Guided Transformer-CNN Framework for Face Super-Resolution
    Qi, Haoran
    Qiu, Yuwei
    Luo, Xing
    Jin, Zhi
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 1589 - 1599
  • [4] Interactformer: Interactive Transformer and CNN for Hyperspectral Image Super-Resolution
    Liu, Yaoting
    Hu, Jianwen
    Kang, Xudong
    Luo, Jing
    Fan, Shaosheng
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [5] Efficient Swin Transformer for Remote Sensing Image Super-Resolution
    Kang, Xudong
    Duan, Puhong
    Li, Jier
    Li, Shutao
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 6367 - 6379
  • [6] Efficient Dual Attention Transformer for Image Super-Resolution
    Park, Soobin
    Jeong, Yuna
    Choi, Yong Suk
    39TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, SAC 2024, 2024, : 963 - 970
  • [7] A Dual CNN for Image Super-Resolution
    Song, Jiagang
    Xiao, Jingyu
    Tian, Chunwei
    Hu, Yuxuan
    You, Lei
    Zhang, Shichao
    ELECTRONICS, 2022, 11 (05)
  • [8] Efficient image super-resolution based on transformer with bidirectional interaction
    Gendy, Garas
    He, Guanghui
    Sabor, Nabil
    APPLIED SOFT COMPUTING, 2024, 165
  • [9] A hybrid of transformer and CNN for efficient single image super-resolution via multi-level distillation
    Zhou, Zhenting
    Li, Guoping
    Wang, Guozhong
    DISPLAYS, 2023, 76
  • [10] Lightweight Wavelet-Based Transformer for Image Super-Resolution
    Ran, Jinye
    Zhang, Zili
    PRICAI 2022: TRENDS IN ARTIFICIAL INTELLIGENCE, PT III, 2022, 13631 : 368 - 382