Cross-Scale KNN Image Transformer for Image Restoration

被引:3
|
作者
Lee, Hunsang [1 ]
Choi, Hyesong [2 ]
Sohn, Kwanghoon [1 ]
Min, Dongbo [2 ]
机构
[1] Yonsei Univ, Sch Elect & Elect Engn, Seoul, South Korea
[2] Ewha Womans Univ, Dept Comp Sci & Engn, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
Image restoration; Transformers; Noise reduction; Complexity theory; Computer vision; Convolutional neural networks; Feature extraction; denoising; deblurring; deraining; transformer; self-attention; k-nn search; low-level vision; ALGORITHMS; NETWORK;
D O I
10.1109/ACCESS.2023.3242556
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Numerous image restoration approaches have been proposed based on attention mechanism, achieving superior performance to convolutional neural networks (CNNs) based counterparts. However, they do not leverage the attention model in a form fully suited to the image restoration tasks. In this paper, we propose an image restoration network with a novel attention mechanism, called cross-scale $k$ -NN image Transformer (CS-KiT), that effectively considers several factors such as locality, non-locality, and cross-scale aggregation, which are essential to image restoration. To achieve locality and non-locality, the CS-KiT builds $k$ -nearest neighbor relation of local patches and aggregates similar patches through local attention. To induce cross-scale aggregation, we ensure that each local patch embraces different scale information with scale-aware patch embedding (SPE) which predicts an input patch scale through a combination of multi-scale convolution branches. We show the effectiveness of the CS-KiT with experimental results, outperforming state-of-the-art restoration approaches on image denoising, deblurring, and deraining benchmarks.
引用
收藏
页码:13013 / 13027
页数:15
相关论文
共 50 条
  • [1] CSformer: Cross-Scale Features Fusion Based Transformer for Image Denoising
    Yin, Haitao
    Ma, Siyuan
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 1809 - 1813
  • [2] Cross-Scale Fusion Transformer for Histopathological Image Classification
    Huang, Sheng-Kai
    Yu, Yu-Ting
    Huang, Chun-Rong
    Cheng, Hsiu-Chi
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2024, 28 (01) : 297 - 308
  • [3] Cross-scale sampling transformer for semantic image segmentation
    Ma, Yizhe
    Yu, Long
    Lin, Fangjian
    Tian, Shengwei
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 44 (02) : 2895 - 2907
  • [4] CPLFormer: Cross-scale Prototype Learning Transformer for Image Snow Removal
    Chen, Sixiang
    Ye, Tian
    Liu, Yun
    Bai, Jinbin
    Chen, Haoyu
    Lin, Yunlong
    Shi, Jun
    Chen, Erkang
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 4228 - 4239
  • [5] Remote sensing image super-resolution via cross-scale hierarchical transformer
    Xiao, Yi
    Yuan, Qiangqiang
    He, Jiang
    Zhang, Liangpei
    GEO-SPATIAL INFORMATION SCIENCE, 2024, 27 (06) : 1914 - 1930
  • [6] Latent Diffusion Enhanced Rectangle Transformer for Hyperspectral Image Restoration
    Li, Miaoyu
    Fu, Ying
    Zhang, Tao
    Liu, Ji
    Dou, Dejing
    Yan, Chenggang
    Zhang, Yulun
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (01) : 549 - 564
  • [7] Deep Multi-Scale Transformer for Remote Sensing Image Restoration
    Li, Yanting
    2024 5TH INTERNATIONAL CONFERENCE ON GEOLOGY, MAPPING AND REMOTE SENSING, ICGMRS 2024, 2024, : 138 - 142
  • [8] Blind Image Quality Index With Cross-Domain Interaction and Cross-Scale Integration
    Hu, Bo
    Zhu, Guang
    Li, Leida
    Gan, Ji
    Li, Weisheng
    Gao, Xinbo
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 2729 - 2739
  • [9] Decomformer: Decompose Self-Attention of Transformer for Efficient Image Restoration
    Lee, Eunho
    Hwang, Youngbae
    IEEE ACCESS, 2024, 12 : 38672 - 38684
  • [10] Cross-Scale Residual Network: A General Framework for Image Super-Resolution, Denoising, and Deblocking
    Zhou, Yuan
    Du, Xiaoting
    Wang, Mingfei
    Huo, Shuwei
    Zhang, Yeda
    Kung, Sun-Yuan
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (07) : 5855 - 5867