Shape completion with azimuthal rotations using spherical gidding-based invariant and equivariant network

被引:0
作者
Wu H. [1 ]
Miao Y. [1 ]
Fu R. [1 ]
机构
[1] School of Mechanical Engineering, Shanghai Jiao Tong University, 800 Dong Chuan Road, Shanghai
基金
中国国家自然科学基金;
关键词
3D vision; Azimuthal rotation; Deep learning; Point cloud completion;
D O I
10.1007/s00521-024-09712-z
中图分类号
学科分类号
摘要
Point cloud completion aims to restore full shapes of objects from their partial views obtained by 3D optical scanners. In order to make point cloud completion become more robust to azimuthal rotations and more adaptive to real-world scenarios, we propose a novel network for simultaneous rotation invariant and equivariant completion with no need of data augmentation, while other existing approaches require separately trained models for different completion types. Our method includes several main steps: First, Density Compensation Mapping (DCM) as well as Aggregative Gaussian Gridding (AGG) modules are introduced to transfer partial point clouds to spherical signals and avoid unbalanced sampling. Second, an encoder based on group correlation is designed to extract rotation invariant global features and equivariant azimuthal features from spherical signals. Third, parallel groups of decoders are proposed to realize rotation invariant completion based on feature fusion. Finally, a feature remapping module as well as Pose Voting Alignment (PVA) algorithm are proposed to unify feature space and realize rotation equivariant completion. Based on these modules, we find that the application of group correlation can be extended to the domain of shape completion; equivariant and invariant completions can be unified in one pipeline, and our inherent rotation equivariant and invariant framework can achieve competitive performances when comparing with existing representative methods. © The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2024.
引用
收藏
页码:13269 / 13292
页数:23
相关论文
共 71 条
[1]  
Song S., Yu F., Zeng A., Chang A.X., Savva M., Funkhouser T., Semantic scene completion from a single depth image, 2017 IEEE conference on computer vision and pattern recognition (CVPR), pp. 190-198, (2017)
[2]  
Varley J., Dechant C., Richardson A., Ruales J., Allen P., Shape completion enabled robotic grasping, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2442-2447, (2017)
[3]  
Charles R.Q., Su H., Kaichun M., Guibas L.J., PointNet: deep learning on point sets for 3D classification and segmentation, 2017 IEEE conference on computer vision and pattern recognition (CVPR), pp. 77-85, (2017)
[4]  
Qi C.R., Yi L., Su H., Guibas L., PointNet++: Deep hierarchical feature learning on point sets in a metric space, Adv Neural Inf Process Syst, (2017)
[5]  
Li Y., Bu R., Sun M., Wu W., Di X., Chen B., PointCNN: Convolution on X-transformed points, Adv Neural Inf Process Syst, (2018)
[6]  
Achlioptas P., Diamanti O., Mitliagkas I., Guibas L., Learning representations and generative models for 3D point clouds, 2018 Proceedings of the 35Th International Conference on Machine Learning (ICML), pp. 40-49, (2018)
[7]  
Groueix T., Fisher M., Kim V.G., Russell B.C., Aubry M., A papier-mache approach to learning 3D surface generation, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 216-224, (2018)
[8]  
Yang Y., Feng C., Shen Y., Tian D., FoldingNet: Point cloud auto-encoder via deep grid deformation, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 206-215, (2018)
[9]  
Yuan W., Khot T., Held D., Mertz C., Hebert M., PCN: Point completion network, 2018 International Conference on 3D Vision (3DV), pp. 728-737, (2018)
[10]  
Tchapmi L.P., Kosaraju V., Rezatofighi H., Reid I., Savarese S., TopNet: structural point cloud decoder, 2019 IEEE/CVF conference on computer vision and pattern recognition (CVPR), pp. 383-392, (2019)