Shape completion with azimuthal rotations using spherical gidding-based invariant and equivariant network

被引:0
作者
Wu H. [1 ]
Miao Y. [1 ]
Fu R. [1 ]
机构
[1] School of Mechanical Engineering, Shanghai Jiao Tong University, 800 Dong Chuan Road, Shanghai
基金
中国国家自然科学基金;
关键词
3D vision; Azimuthal rotation; Deep learning; Point cloud completion;
D O I
10.1007/s00521-024-09712-z
中图分类号
学科分类号
摘要
Point cloud completion aims to restore full shapes of objects from their partial views obtained by 3D optical scanners. In order to make point cloud completion become more robust to azimuthal rotations and more adaptive to real-world scenarios, we propose a novel network for simultaneous rotation invariant and equivariant completion with no need of data augmentation, while other existing approaches require separately trained models for different completion types. Our method includes several main steps: First, Density Compensation Mapping (DCM) as well as Aggregative Gaussian Gridding (AGG) modules are introduced to transfer partial point clouds to spherical signals and avoid unbalanced sampling. Second, an encoder based on group correlation is designed to extract rotation invariant global features and equivariant azimuthal features from spherical signals. Third, parallel groups of decoders are proposed to realize rotation invariant completion based on feature fusion. Finally, a feature remapping module as well as Pose Voting Alignment (PVA) algorithm are proposed to unify feature space and realize rotation equivariant completion. Based on these modules, we find that the application of group correlation can be extended to the domain of shape completion; equivariant and invariant completions can be unified in one pipeline, and our inherent rotation equivariant and invariant framework can achieve competitive performances when comparing with existing representative methods. © The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2024.
引用
收藏
页码:13269 / 13292
页数:23
相关论文
共 71 条
[31]  
Aoki Y., Goforth H., Srivatsan R.A., Lucey S., PointNetLK: Robust efficient point cloud registration using PointNet, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7156-7165, (2019)
[32]  
Yi L., Zhao W., Wang H., Sung M., Guibas L.J., GSPN: generative shape proposal network for 3D instance segmentation in point cloud, 2019 IEEE/CVF conference on computer vision and pattern recognition (CVPR), pp. 3942-3951, (2019)
[33]  
Yang B., Wang J., Clark R., Hu Q., Wang S., Markham A., Trigoni N., Learning object bounding boxes for 3D instance segmentation on point clouds, Adv Neural Inf Process Syst, (2019)
[34]  
Hu Q., Yang B., Xie L., Rosa S., Guo Y., Wang Z., Trigoni N., Markham A., RandLA-Net: Efficient semantic segmentation of large-scale point clouds, IEEE/CVF Conference on Computer Vision and Pattern Recognition, (2020)
[35]  
Qi C.R., Litany O., He K., Guibas L., Deep hough voting for 3D object detection in point clouds, 2019 IEEE/CVF international conference on computer vision (ICCV), pp. 9276-9285, (2019)
[36]  
Shi S., Wang X., Li H., PointRCNN: 3D object proposal generation and detection from point cloud, 2019 IEEE/CVF conference on computer vision and pattern recognition (CVPR), pp. 770-779, (2019)
[37]  
Hirschmuller H., Accurate and efficient stereo processing by semi-global matching and mutual information, IEEE Computer Society Conference on Computer Vision and Pattern Recognition, (2005)
[38]  
Yao Y.Y., Luo Z., Li L.S., Fang F.T., Quan L., MVSNet: Depth inference for unstructured multi-view stereo, European Conference on Computer Vision, (2018)
[39]  
Chen R., Han S., Xu J., Su H., Point-based multi-view stereo network, 2019 IEEE/CVF international conference on computer vision (ICCV), pp. 1538-1547, (2019)
[40]  
Wei Y., Liu S., Wang Z., Lu J., Conditional single-view shape generation for multi-view stereo reconstruction, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9643-9652, (2019)