Magic ELF: Image Deraining Meets Association Learning and Transformer

被引:15
|
作者
Jiang, Kui [1 ]
Wang, Zhongyuan [1 ]
Chen, Chen [2 ]
Wang, Zheng [1 ]
Cui, Laizhong [3 ]
Lin, Chia-Wen [4 ]
机构
[1] Wuhan Univ, NERCMS, Wuhan, Peoples R China
[2] Univ Cent Florida, CRCV, Orlando, FL 32816 USA
[3] Shenzhen Univ, Hangzhou, Peoples R China
[4] Natl Tsing Hua Univ, Hsinchu, Taiwan
基金
中国国家自然科学基金;
关键词
Image Deraining; Self-attention; Association Learning; QUALITY ASSESSMENT; RAIN REMOVAL; NETWORK;
D O I
10.1185/3503161.3547760
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Convolutional neural network (CNN) and Transformer have achieved great success in multimedia applications. However, little effort has been made to effectively and efficiently harmonize these two architectures to satisfy image deraining. This paper aims to unify these two architectures to take advantage of their learning merits for image deraining. In particular, the local connectivity and translation equivariance of CNN and the global aggregation ability of self-attention (SA) in Transformer are fully exploited for specific local context and global structure representations. Based on the observation that rain distribution reveals the degradation location and degree, we introduce degradation prior to help background recovery and accordingly present the association refinement deraining scheme. A novel multi-input attention module (MAM) is proposed to associate rain perturbation removal and background recovery. Moreover, we equip our model with effective depth-wise separable convolutions to learn the specific feature representations and trade off computational complexity. Extensive experiments show that our proposed method (dubbed as ELF) outperforms the state-of-the-art approach (MPRNet) by 0.25 dB on average, but only accounts for 11.7% and 42.1% of its computational cost and parameters.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Two-stage learning framework for single image deraining
    Jiang, Rui
    Li, Yaoshun
    Chen, Cheng
    Liu, Wei
    IET IMAGE PROCESSING, 2023, 17 (05) : 1449 - 1463
  • [42] Context-Enhanced Representation Learning for Single Image Deraining
    Wang, Guoqing
    Sun, Changming
    Sowmya, Arcot
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2021, 129 (05) : 1650 - 1674
  • [43] Spatial-guided informative semantic joint transformer for single-image deraining
    Li, Haiyan
    Peng, Shaolin
    Lang, Xun
    Ye, Shuhua
    Li, Hongsong
    JOURNAL OF SUPERCOMPUTING, 2024, 80 (05): : 6522 - 6551
  • [44] Memory Oriented Transfer Learning for Semi-Supervised Image Deraining
    Huang, Huaibo
    Yu, Aijing
    He, Ran
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 7728 - 7737
  • [45] Exploring high-quality image deraining Transformer via effective large kernel attention
    Dong, Haobo
    Song, Tianyu
    Qi, Xuanyu
    Jin, Jiyu
    Jin, Guiyue
    Fan, Lei
    VISUAL COMPUTER, 2025, 41 (04): : 2545 - 2561
  • [46] UC-former: A multi-scale image deraining network using enhanced transformer
    Zhou, Weina
    Ye, Linhui
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2024, 248
  • [47] Continuous learning deraining network based on residual FFT convolution and contextual transformer module
    Zhang, Zhijia
    Wu, Sinan
    Peng, Xinming
    Wang, Wanting
    Li, Rui
    IET IMAGE PROCESSING, 2023, 17 (03) : 747 - 760
  • [48] LViT: Language Meets Vision Transformer in Medical Image Segmentation
    Li, Zihan
    Li, Yunxiang
    Li, Qingde
    Wang, Puyang
    Guo, Dazhou
    Lu, Le
    Jin, Dakai
    Zhang, You
    Hong, Qingqi
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2024, 43 (01) : 96 - 107
  • [49] When Fast Fourier Transform Meets Transformer for Image Restoration
    Jiang, Xingyu
    Zhang, Xiuhui
    Gao, Ning
    Deng, Yue
    COMPUTER VISION - ECCV 2024, PT XLV, 2025, 15103 : 381 - 402
  • [50] Exploring high-quality image deraining Transformer via effective large kernel attentionExploring high-quality image deraining Transformer via effective large kernel attentionH. Dong et al.
    Haobo Dong
    Tianyu Song
    Xuanyu Qi
    Jiyu Jin
    Guiyue Jin
    Lei Fan
    The Visual Computer, 2025, 41 (4) : 2545 - 2561