VResNet: A Deep Learning Architecture for Image Inpainting of Irregular Damaged Images

被引:0
作者
Sharma S. [1 ]
Rani R. [1 ]
机构
[1] Department of Computer Science and Engineering, Dr. B. R. Ambedkar National Institute of Technology, Jalandhar
关键词
Deep learning; Image inpainting; Irregular mask; Residual learning; U-net;
D O I
10.1007/s42979-023-02523-4
中图分类号
学科分类号
摘要
In computer vision, image inpainting is a famous problem to automatically reconstruct the damaged part of the image according to the undamaged portion of an image. Inpainting irregular damaged areas in the image is still challenging. Deep learning-based techniques have given us a fantastic performance over the last few years. In this paper, we propose VResNet, a deep-learning approach for image inpainting, inspired by U-Net architecture and the residual framework. Since deeper neural networks are extra hard to train, the superficial convolution block in U-Net architecture is replaced by the residual learning block in the proposed approach to simplify the training of deeper neural networks. To develop an effective and adaptable model, an extensive series of experiments was conducted using the Paris-Street-View dataset. Our proposed method achieved notable results, including a PSNR of 20.65, an SSIM of 0.65, an L1 Loss of 6.90, and a total loss (L Total) of 0.30 on the Paris-Street-View dataset. These outcomes clearly demonstrate the superior performance of our model when compared to other techniques. The paper presents both qualitative and quantitative comparisons to provide a comprehensive assessment of our approach. © 2024, The Author(s), under exclusive licence to Springer Nature Singapore Pte Ltd.
引用
收藏
相关论文
共 51 条
  • [1] Ballester C., Bertalmio M., Caselles V., Sapiro G., Verdera J., Filling-in by joint interpolation of vector fields and gray levels, IEEE Trans Image Process, 10, 8, pp. 1200-1211, (2001)
  • [2] Barnes C., Shechtman E., Finkelstein A., Goldman D.B., Patchmatch: A randomized correspondence algorithm for structural image editing, ACM Trans. Graph., 28, 3, (2009)
  • [3] Bertalmio M., Sapiro G., Caselles V., Ballester C., Image inpainting, Proceedings of the 27Th Annual Conference on Computer Graphics and Interactive Techniques, pp. 417-424, (2000)
  • [4] Bishop C.M., Et al., Neural networks for pattern recognition, (1995)
  • [5] Chen B., Qi X., Zhou Y., Yang G., Zheng Y., Xiao B., Image splicing localization using residual image and residual-based fully convolutional network, J Vis Commun Image Represent, 73, (2020)
  • [6] El Biach F.Z., Iala I., Laanaya H., Minaoui K., Encoder-decoder based convolutional neural networks for image forgery detection, Multimedia Tools and Applications, pp. 1-18, (2021)
  • [7] Haritha L., Prajith C.A., Image inpainting using deep learning techniques: A review, 2023 International Conference on Control, Communication and Computing (ICCC), pp. 1-6, (2023)
  • [8] Hays J., Efros A.A., Scene completion using millions of photographs, Commun ACM, 51, 10, pp. 87-94, (2008)
  • [9] He K., Zhang X., Ren S., Sun J., Delving deep into rectifiers: Surpassing human-level performance on imagenet classification, Proceedings of the IEEE International Conference on Computer Vision, pp. 1026-1034, (2015)
  • [10] He K., Zhang X., Ren S., Sun J., Spatial pyramid pooling in deep convolutional networks for visual recognition, IEEE Trans Pattern Anal Mach Intell, 37, 9, pp. 1904-1916, (2015)