SFI-Swin: symmetric face inpainting with swin transformer by distinctly learning face components distributions

被引:1
作者
Givkashi M. [1 ]
Naderi M. [1 ]
Karimi N. [1 ]
Shirani S. [2 ]
Samavi S. [2 ,3 ]
机构
[1] Department of Electrical and Computer Engineering, Isfahan University of Technology, Isfahan
[2] Department of Electrical and Computer Engineering, McMaster University, Hamilton
[3] Computer Science Department, Seattle University, Seattle
关键词
Deep Learning; Face Inpainting; Swin Transformer;
D O I
10.1007/s11042-024-19365-8
中图分类号
TB18 [人体工程学]; Q98 [人类学];
学科分类号
030303 ; 1201 ;
摘要
Image inpainting consists of filling holes or missing parts of an image. Inpainting face images with symmetric characteristics is more challenging than inpainting a natural scene. None of the powerful existing models can fill out the missing parts of an image while considering the symmetry and homogeneity of the picture. Moreover, the metrics that assess a repaired face image quality cannot measure the preservation of symmetry between the rebuilt and existing parts of a face. In this paper, we intend to solve the symmetry problem in the face inpainting task by using multiple discriminators that check each face organ’s reality separately and a transformer-based network. We also propose "symmetry concentration score" as a new metric for measuring the symmetry of a repaired face image. The quantitative and qualitative results show the superiority of our proposed method compared to some of the recently proposed algorithms in terms of the reality, symmetry, and homogeneity of the inpainted parts. The code for the proposed method is available at https://github.com/mohammadrezanaderi4/SFI-Swin. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2024.
引用
收藏
页码:17581 / 17595
页数:14
相关论文
共 38 条
[1]  
Zhao S., Et al., Large scale image completion via co-modulated generative adversarial networks, Arxiv Prepr:Arxiv, (2021)
[2]  
Zeng Y., Fu J., Chao H., Guo B., Aggregated contextual transformations for high-resolution image inpainting, IEEE Trans Vis Comput Graph, 29, 7, pp. 3266-3280, (2022)
[3]  
Ma Y., Et al., Region-wise generative adversarial image inpainting for large missing areas, IEEE Trans Cybern, 53, 8, pp. 5226-5239, (2022)
[4]  
Nazeri K., Ng E., Joseph T., Qureshi F.Z., Ebrahimi M., Edgeconnect:Generative image inpainting with adversarial edge learning, Arxiv Prepr:Arxiv, (2019)
[5]  
Yu J., Lin Z., Yang J., Shen X., Lu X., Huang T.S., Free-form image inpainting with gated convolution, Proceedings of the IEEE Conference on Computer Vision, pp. 4471-4480, (2019)
[6]  
Suvorov R., Et al., Resolution-robust large mask inpainting with Fourier convolutions, Proceedings IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 2149-2159, (2022)
[7]  
Isola P., Zhu J.Y., Zhou T., Efros A.A., Image-to-image translation with conditional adversarial networks, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1125-1134, (2017)
[8]  
Liu Z., Et al., Swin transformer: Hierarchical vision transformer using shifted windows, Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 10012-10022, (2021)
[9]  
Zhang X., Zhai D., Li T., Zhou Y., Lin Y., Image inpainting based on deep learning: a review, Inf Fusion, 90, pp. 74-94, (2022)
[10]  
Qin Z., Zeng Q., Zong Y., Xu F., Image inpainting based on deep learning: A review, Displays, 69, (2021)