Multi-guided feature refinement for point cloud semantic segmentation with weakly supervision

被引:0
作者
Wang, Yufan [1 ]
Zhao, Qunfei [1 ]
Xia, Zeyang [2 ]
机构
[1] Shanghai Jiao Tong Univ, Inst Image Proc & Pattern Recognit, Dept Automat, Shanghai 200240, Peoples R China
[2] Shanghai Jiao Tong Univ, Sch Mech Engn, Shanghai 200240, Peoples R China
关键词
Point cloud; Semantic segmentation; Weakly-supervision learning; Feature refinement;
D O I
10.1016/j.knosys.2025.113050
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Point cloud segmentation is a widely concerned task for 3D part/scene parsing and many learning-based methods are proposed to improve the performance significantly. However, the performance is limited by the quality and quantity of labeled data. Therefore, we propose a multi-guided feature refinement (MGFR) to capture more effective representation with fewer annotations. Specifically, MGFR is a point-wise method based on a hybrid neighbor system and consists of feature aggregation and weight refinement. Feature aggregation is implemented in an attention-based manner guided by explicit information (structure geometry prior and RGB prior), neighbor information and prototype information. Weight refinement is a probabilistic method which is guided by the effective components of prototype extracted from neighbor members. The refined point feature of MGFR is provided with more local smoothness and global consistency, which can improve the performance on different instances of the same class and reduce the counterintuitive error around classification boundary or isolated outliers. Furthermore, we also use a neighbor-based contrastive loss, a prototype-based loss with regularization and a neighbor-based multiple instance loss to achieve local optimization and regularize the distribution of point embedding. Experimentally, we evaluate MGFR on ShapeNet Part dataset, Stanford 2D-3D (S3DIS) dataset and ScanNet, which shows the effectiveness in weakly-supervised task.
引用
收藏
页数:12
相关论文
共 66 条
  • [51] He S., Jiang X., Jiang W., Ding H., Prototype adaption and projection for few-and zero-shot 3d point cloud semantic segmentation, IEEE Trans. Image Process., 32, pp. 3199-3211, (2023)
  • [52] Zhang Y., Qu Y., Xie Y., Li Z., Zheng S., Li C., Perturbed self-distillation: Weakly supervised large-scale point cloud semantic segmentation, pp. 15520-15528, (2021)
  • [53] Li M., Xie Y., Shen Y., Ke B., Qiao R., Ren B., Lin S., Ma L., Hybridcr: Weakly-supervised 3d point cloud semantic segmentation via hybrid contrastive regularization, pp. 14930-14939, (2022)
  • [54] Yang C.-K., Wu J.-J., Chen K.-S., Chuang Y.-Y., Lin Y.-Y., An mil-derived transformer for weakly supervised point cloud segmentation, pp. 11830-11839, (2022)
  • [55] He K., Fan H., Wu Y., Xie S., Girshick R., Momentum Contrast for Unsupervised Visual Representation Learning, pp. 9726-9735, (2020)
  • [56] Chang A.X., Funkhouser T., Guibas L., Hanrahan P., Huang Q., Li Z., Savarese S., Savva M., Song S., Su H., ShapeNet: An information-rich 3D model repository, Comput. Sci., (2015)
  • [57] Armeni I., Sax S., Zamir A.R., Savarese S., Joint 2d-3d-semantic data for indoor scene understanding, (2017)
  • [58] Wang Z., Yu X., Rao Y., Zhou J., Lu J., Take-a-photo: 3d-to-2d generative pre-training of point cloud models, pp. 5640-5650, (2023)
  • [59] Laine S., Aila T., Temporal ensembling for semi-supervised learning, (2016)
  • [60] Tarvainen A., Valpola H., Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results, Adv. Neural Inf. Process. Syst., 30, (2017)