Multi-scale network for single image deblurring based on ensemble learning module

被引:0
作者
Wu W. [1 ,2 ]
Pan Y. [1 ,2 ]
Su N. [1 ]
Wang J. [1 ,2 ]
Wu S. [3 ]
Xu Z. [4 ]
Yu Y. [5 ]
Liu Y. [6 ]
机构
[1] State Key Laboratory of Media Convergence and Communication, Communication University of China, Beijing
[2] School of Computer and Cyberspace Security, Communication University of China, Beijing
[3] School of Electronics and Information Engineering, Harbin Institute of Technology, Harbin
[4] College of Computer and Control Engineering, Northeast Forestry University, Harbin
[5] School of Computer and Information Engineering, Tianjin Chengjian University, Tianjin
[6] School of Computer Science and Technology, Anhui University, Hefei
基金
中国国家自然科学基金;
关键词
Ensemble learning; Image deblurring; Multi-scale; Neural networks;
D O I
10.1007/s11042-024-19295-5
中图分类号
TN911 [通信理论];
学科分类号
081002 ;
摘要
In this paper, we have identified two primary issues with current multi-scale image deblurring methods. On the one hand, the blurring scale is ignored. On the other hand, the context information of images is not fully utilized, and the spatial detail information of images is lost. To better solve the above problems, on top of that, this paper proposes a multi-scale network for single image deblurring. The designed network contains two parts: 1) Ensemble Learning Model: The Weak Learning Model in this module connects multiple weak learners in series to increase network depth. Moreover, through multi-scale feature ensemble module, the output features and intermediate features of different weak learner are integrated, and finally strong learner is used to enhance the features with semantic information at different scales; 2) Spatial Feature Enhancement Module: This module can preserve spatial detail information of images and compensate for the disadvantages of losing image spatial details due to continuous downsampling in the Ensemble Learning Model structure. Our proposed deblurring model has been extensively evaluated on several benchmark datasets, and the model achieves superior performance compared to the most advanced deblurring methods. In particular, the peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) values are 33.57 dB and 0.965 on the GoPro dataset, respectively, with PSNR and SSIM improving by 0.49 dB and 0.002 over NAFNet, and 0.65 dB and 0.004 over Restormer. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2024.
引用
收藏
页码:9045 / 9064
页数:19
相关论文
共 71 条
[1]  
Chan T.F., Wong C.-K., Total variation blind deconvolution, IEEE Trans Image Process, 7, 3, pp. 370-375, (1998)
[2]  
Goldstein A., Fattal R., Blur-Kernel Estimation from Spectral Irregularities., Part V, 12, pp. 622-635, (2012)
[3]  
Hyun Kim T., Ahn B., Mu Lee K., Dynamic scene deblurring, Proceedings of the IEEE International Conference on Computer Vision, pp. 3160-3167, (2013)
[4]  
Krishnan D., Tay T., Fergus R., Blind deconvolution using a normalized sparsity measure, CVPR, pp. 233-240, (2011)
[5]  
Pan J., Hu Z., Su Z., Yang M.-H., Deblurring face images with exemplars, Computer vision–ECCV 2014: 13Th European Conference, pp. 47-62, (2014)
[6]  
Shan Q., Jia J., Agarwala A., High-quality motion deblurring from a single image, Acm Trans Graph (TOG), 27, 3, pp. 1-10, (2008)
[7]  
Xu L., Jia J., Two-phase kernel estimation for robust motion deblurring, Computer vision–ECCV 2010: 11Th European Conference on Computer Vision, pp. 157-170, (2010)
[8]  
Xu L., Zheng S., Jia J., Unnatural l0 sparse representation for natural image deblurring, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1107-1114, (2013)
[9]  
Zoran D., Weiss Y., From learning models of natural image patches to whole image restoration, 2011 International Conference on Computer Vision, pp. 479-486, (2011)
[10]  
Nah S., Hyun Kim T., Mu Lee K., Deep multi-scale convolutional neural network for dynamic scene deblurring, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3883-3891, (2017)