Image Inpainting Based on Structural Constraint and Multi-Scale Feature Fusion

被引:4
作者
Fan, Yao [1 ]
Shi, Yingnan [1 ]
Zhang, Ningjun [2 ]
Chu, Yanli [3 ]
机构
[1] Xizang Minzu Univ, Coll Informat Engn, Xianyang 712082, Shaanxi, Peoples R China
[2] Zhengzhou Inst Sci & Technol, Coll Informat Engn, Zhengzhou 450052, Peoples R China
[3] Univ CAPF, Coll Equipment Management & Guarantee, Xian 712000, Peoples R China
来源
IEEE ACCESS | 2023年 / 11卷
基金
中国国家自然科学基金;
关键词
Deep learning; image inpainting; edge repair; dilated residual feature pyramid fusion; dilated multi-scale attention fusion; TEXTURE SYNTHESIS; OBJECT REMOVAL;
D O I
10.1109/ACCESS.2023.3246062
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
When repairing masked images based on deep learning, there is usually insufficient representation of multi-level information and inadequate utilization of long distance features. To solve the problems, this paper proposes a second-order generative image inpainting model based on Structural Constraints and Multi-scale Feature Fusion (SCMFF). The SCMFF model consists of two parts: edge repair network and image inpainting network. The edge repair network combines the auto-encoder with the Dilated Residual Feature Pyramid Fusion (DRFPF) module, which improves the representation of multi-level semantic information and structural details of images, thus achieves better edge repair. Then, the image inpainting network embeds the Dilated Multi-scale Attention Fusion (DMAF) module in the auto-encoder for texture synthesis with the real edge as the prior condition, and achieves fine-grained inpainting under the edge constraint by aggregating the long-distance features of different dimensions. Finally, the edge repair results are used to replace the real edge, and the two networks are fused and trained to achieve end-to-end repair from the masked image to the complete image. The model is compared with the advanced methods on datasets including Celeba, Facade and Places2. The quantitative results show that the four metrics of LPIPS, MAE, PSNR and SSIM are improved by 0.0124-0.0211, 3.787-6.829, 2.934dB-5.730dB and 0.034-0.132, respectively. The qualitative results show that the edge distribution in the center of the hole reconstructed by the SCMFF model is more uniform, and the texture synthesis effect is more in line with human visual perception.
引用
收藏
页码:16567 / 16587
页数:21
相关论文
共 43 条
  • [1] Filling-in by joint interpolation of vector fields and gray levels
    Ballester, C
    Bertalmio, M
    Caselles, V
    Sapiro, G
    Verdera, J
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2001, 10 (08) : 1200 - 1211
  • [2] Cheng W., 2005, P INT C COMPUTER GRA, P64, DOI DOI 10.5120/17692-8659
  • [3] Region filling and object removal by exemplar-based image inpainting
    Criminisi, A
    Pérez, P
    Toyama, K
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2004, 13 (09) : 1200 - 1212
  • [4] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
  • [5] Image Inpainting Using Nonlocal Texture Matching and Nonlinear Filtering
    Ding, Ding
    Ram, Sundaresh
    Rodriguez, Jeffrey J.
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (04) : 1705 - 1719
  • [6] Efros AA, 2001, COMP GRAPH, P341, DOI 10.1145/383259.383296
  • [7] Image Style Transfer Using Convolutional Neural Networks
    Gatys, Leon A.
    Ecker, Alexander S.
    Bethge, Matthias
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 2414 - 2423
  • [8] Goodfellow I., 2014, PROC ADV NEURAL INF, P2676
  • [9] Image Inpainting via Conditional Texture and Structure Dual Generation
    Guo, Xiefan
    Yang, Hongyu
    Huang, Di
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 14114 - 14123
  • [10] Hongyu Liu, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12347), P725, DOI 10.1007/978-3-030-58536-5_43