Self-supervised reflectance-guided 3d shape reconstruction from single-view images

被引:1
作者
Fang, Binbin [1 ]
Xiao, Nanfeng [1 ]
机构
[1] South China Univ Technol, Sch Comp Sci & Engn, Guangzhou 510006, Peoples R China
关键词
3D reconstruction; Self-supervised; Reflectance; Single-view images; FACE RECONSTRUCTION;
D O I
10.1007/s10489-022-03724-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
3D shape reconstruction from a single-view image is an utterly ill-posed and challenging problem, while multi-view methods can reconstruct an object's shape only from raw images. However, these raw images should be shot in a static scene, to promise that corresponding features in the images can be mapped to the same spatial location. Recent single-view methods need only single-view images of static or dynamic objects, by turning to prior knowledge to mine the latent multi-view information in single-view images. Some of them utilize prior models (e.g. rendering-based or style-transfer-based) to generate novel-view images, which are however not sufficiently accurate, to feed their model. In this paper, we represent Augmented Self-Supervised 3D Reconstruction with Monotonous Material (ASRMM) approach, trained end-to-end in a self-supervised manner, to obtain the 3D reconstruction of a category-specific object, without any relevant prior models for novel-view images. Our approach draws inspiration from the experience that (1) high quality multi-view images are difficult to obtain, and (2) the shape of an object of single material can be visually inferred more easily, rather than of multiple kinds of complex material. As to practice these motivations, ASRMM makes material monotonous in its diffuse part by setting reflectance an identical value, and apply this idea on the source and reconstruction images. Experiments show that our model can reasonably reconstruct the 3D model of faces, cats, cars and birds from their collections of single-view images, and the experiments also show that our approach can be generalized to different reconstruction tasks, including unsupervised depth-based reconstruction and 2D supervised mesh reconstruction, and achieve promising improvement in the quality of the reconstructed shape and the texture.
引用
收藏
页码:6966 / 6977
页数:12
相关论文
共 37 条
[1]  
[Anonymous], 2016, PROC CVPR IEEE, DOI DOI 10.1109/CVPR.2016.445
[2]  
[Anonymous], 1760, Photometria Sive De Mensura Et Gradibus Luminis, Colorum Et Umbrae
[3]  
Blinn J.F., 1977, P 4 ANN C COMPUTER G, P192, DOI DOI 10.1145/563858.563893
[4]  
Chang AX, 2015, ARXIV 151203012
[5]  
Chen T, 2020, PR MACH LEARN RES, V119
[6]  
Chen Wenzheng, 2019, Advances in neural information processing systems, P9609
[7]  
Cook R. L., 1982, ACM Trans Graph, V1, P7, DOI [10.1145/357290.357293, DOI 10.1145/357290.357293]
[8]   Accurate 3D Face Reconstruction with Weakly-Supervised Learning: From Single Image to Image Set [J].
Deng, Yu ;
Yang, Jiaolong ;
Xu, Sicheng ;
Chen, Dong ;
Jia, Yunde ;
Tong, Xin .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, :285-295
[9]   Practical Face Reconstruction via Differentiable Ray Tracing [J].
Dib, A. ;
Bharaj, G. ;
Ahn, J. ;
Thebault, C. ;
Gosselin, P. ;
Romeo, M. ;
Chevallier, L. .
COMPUTER GRAPHICS FORUM, 2021, 40 (02) :153-164
[10]   GANFIT: Generative Adversarial Network Fitting for High Fidelity 3D Face Reconstruction [J].
Gecer, Baris ;
Ploumpis, Stylianos ;
Kotsia, Irene ;
Zafeiriou, Stefanos .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :1155-1164