Frequency-guided Dual Sparse Self-Attention Algorithm for Single Image Deraining

被引:0
|
作者
Wen Y.-B. [1 ]
Gao T. [1 ]
Chen T. [1 ]
Zhang Q.-X. [1 ]
机构
[1] School of Information Engineering, Chang’an University, Shaanxi, Xi’an
来源
基金
中国国家自然科学基金;
关键词
computer vision; frequency-guided learning; image deraining; sparse attention; spatial shift; Transformer;
D O I
10.12263/DZXB.20221420
中图分类号
学科分类号
摘要
Existing Transformer-based algorithms for single image deraining achieve state-of-the-art performance but leading to reasonable computational loads while failing to effectively process real-world rainy images. To this end, we propose a frequency-guided dual sparse self-attention Transformer for single image deraining (FDSATFormer). Initially, our proposed method utilizes the spatial sparse factor and the channel reduction factor to extract accurate global information and significantly decreases the amount of computation. Furthermore, we present dual sparse self-attention feature learning network (DSFL) to deal with the problem that Transformer is difficult to represent self-attention on high-resolution feature maps. Meanwhile, by exploring the spectral changes of rainy image before and after removing rain streaks, we develop a frequency-guided feature enhancement module (FFE), which exploits the global information from the frequency domain to guide the accurate learning of spatial features in network encoders. In addition, the encoder and decoder of most existing methods follow the similar principles, resulting in almost double computational burden. To handle with this issue, we propose a hierarchical feature decoding and reconstructing network (HFDR), which uses non-parametric spatial neighborhood shift (SNS) to construct the feature decoding network and achieves fine results while further reducing the overall computing burden. Experimental results show that, our method improves the average peak signal noise ratio by 3.13 dB and 0.12 dB, and achieves performance gains of 1.39% and 1.07% in average structure similarity over the state-of-the-art Uformer and Restormer. © 2023 Chinese Institute of Electronics. All rights reserved.
引用
收藏
页码:2812 / 2820
页数:8
相关论文
共 34 条
  • [1] GAI S, WANG J S., Image raindrop algorithm research using nonlocal attention enhanced network based on deep learning, Acta Electronica Sinica, 48, 10, pp. 1899-1908, (2020)
  • [2] FU X Y, HUANG J B, DING X H, Et al., Clearing the skies: A deep network architecture for single-image rain removal, IEEE Transactions on Image Processing, 26, 6, pp. 2944-2956, (2017)
  • [3] QIN J, WEN Y B, GAO T, Et al., Transformer-based diffusion model for all-in-one weather-degraded image restoration, Journal of Shanghai Jiaotong University, (2023)
  • [4] KIM J H, LEE C, SIM J Y, Et al., Single-image deraining using an adaptive nonlocal means filter, 2013 IEEE International Conference on Image Processing, pp. 914-917, (2014)
  • [5] XU J, ZHAO W, LIU P, Et al., Removing rain and snow in a single image using guided filter, 2012 IEEE International Conference on Computer Science and Automation Engineering (CSAE), pp. 304-307, (2012)
  • [6] ZHENG X H, LIAO Y H, GUO W, Et al., Single-image-based rain and snow removal using multi-guided filter, Neural Information Processing: 20th International Conference, ICONIP 2013, pp. 258-265, (2013)
  • [7] WEI W, MENG D Y, ZHAO Q A, Et al., Semi-supervised transfer learning for image rain removal, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3872-3881, (2020)
  • [8] ZHANG H, PATEL V M., Density-aware single image deraining using a multi-stream dense network, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 695-704, (2018)
  • [9] YASARLA R, PATEL V M., Uncertainty guided multi-scale residual learning-using a cycle spinning CNN for single image de-raining, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8397-8406, (2020)
  • [10] LI X A, WU J L, LIN Z C, Et al., Recurrent squeeze-and-excitation context aggregation net for single image deraining, Computer Vision ‒ ECCV 2018, pp. 262-277, (2018)