Joint features-guided linear transformer and CNN for efficient image super-resolution

被引:2
|
作者
Wang, Bufan [1 ]
Zhang, Yongjun [1 ]
Long, Wei [1 ]
Cui, Zhongwei [2 ]
机构
[1] Guizhou Univ, Coll Comp Sci & Technol, State Key Lab Publ Big Data, Guiyang 550025, Guizhou, Peoples R China
[2] Guizhou Educ Univ, Sch Math & Big Data, Guiyang 550018, Peoples R China
关键词
Image super-resolution; Multi-level contextual information; Linear self-attention; Lightweight network; NETWORK;
D O I
10.1007/s13042-024-02277-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Integrating convolutional neural networks (CNNs) and transformers has notably improved lightweight single image super-resolution (SISR) tasks. However, existing methods lack the capability to exploit multi-level contextual information, and transformer computations inherently add quadratic complexity. To address these issues, we propose a Joint features-Guided Linear Transformer and CNN Network (JGLTN) for efficient SISR, which is constructed by cascading modules composed of CNN layers and linear transformer layers. Specifically, in the CNN layer, our approach employs an inter-scale feature integration module (IFIM) to extract critical latent information across scales. Then, in the linear transformer layer, we design a joint feature-guided linear attention (JGLA). It jointly considers adjacent and extended regional features, dynamically assigning weights to convolutional kernels for contextual feature selection. This process garners multi-level contextual information, which is used to guide linear attention for effective information interaction. Moreover, we redesign the method of computing feature similarity within the self-attention, reducing its computational complexity to linear. Extensive experiments shows that our proposal outperforms state-of-the-art models while balancing performance and computational costs.
引用
收藏
页码:5765 / 5780
页数:16
相关论文
共 50 条
  • [41] Spatial Transformer Generative Adversarial Network for Image Super-Resolution
    Rempakos, Pantelis
    Vrigkas, Michalis
    Plissiti, Marina E.
    Nikou, Christophoros
    IMAGE ANALYSIS AND PROCESSING, ICIAP 2023, PT I, 2023, 14233 : 399 - 411
  • [42] HSACT: A hierarchical semantic-aware CNN-Transformer for remote sensing image spectral super-resolution
    Zhou, Chengle
    He, Zhi
    Zou, Liwei
    Li, Yunfei
    Plaza, Antonio
    NEUROCOMPUTING, 2025, 636
  • [43] Gradient-Guided Single Image Super-Resolution Based on Joint Trilateral Feature Filtering
    Zuo, Yifan
    Xie, Jiacheng
    Wang, Hao
    Fang, Yuming
    Liu, Deyang
    Wen, Wenying
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (02) : 505 - 520
  • [44] Fully Cross-Attention Transformer for Guided Depth Super-Resolution
    Ariav, Ido
    Cohen, Israel
    SENSORS, 2023, 23 (05)
  • [45] When Handcrafted Filter Meets CNN: A Lightweight Conv-Filter Mixer Network for Efficient Image Super-Resolution
    Wu, Zhijian
    Liu, Wenhui
    Huang, Dingjiang
    PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 722 - 730
  • [46] Separable Modulation Network for Efficient Image Super-Resolution
    Wu, Zhijian
    Li, Jun
    Huang, Dingjiang
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 8086 - 8094
  • [47] A very lightweight and efficient image super-resolution network?
    Gao, Dandan
    Zhou, Dengwen
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 213
  • [48] Adaptive Feature Selection Modulation Network for Efficient Image Super-Resolution
    Wu, Chen
    Wang, Ling
    Su, Xin
    Zheng, Zhuoran
    IEEE SIGNAL PROCESSING LETTERS, 2025, 32 : 1231 - 1235
  • [49] RepECN: Making ConvNets Better Again for Efficient Image Super-Resolution
    Chen, Qiangpu
    Qin, Jinghui
    Wen, Wushao
    SENSORS, 2023, 23 (23)
  • [50] Single Image Super-Resolution Using Lightweight CNN with Maxout Units
    Choi, Jae-Seok
    Kim, Munchurl
    COMPUTER VISION - ACCV 2018, PT VI, 2019, 11366 : 471 - 487