A Review of panoptic segmentation for mobile mapping point clouds

被引:4
作者
Xiang, Binbin [1 ]
Yue, Yuanwen [1 ]
Peters, Torben [1 ]
Schindler, Konrad [1 ]
机构
[1] Swiss Fed Inst Technol, Photogrammetry & Remote Sensing, CH-8093 Zurich, Switzerland
关键词
Mobile mapping point clouds; 3D panoptic segmentation; 3D semantic segmentation; 3D instance segmentation; 3D deep learning backbones; CLASSIFICATION; NETWORKS; DATASET;
D O I
10.1016/j.isprsjprs.2023.08.008
中图分类号
P9 [自然地理学];
学科分类号
0705 ; 070501 ;
摘要
3D point cloud panoptic segmentation is the combined task to (i) assign each point to a semantic class and (ii) separate the points in each class into object instances. Recently there has been an increased interest in such comprehensive 3D scene understanding, building on the rapid advances of semantic segmentation due to the advent of deep 3D neural networks. Yet, to date there is very little work about panoptic segmentation of outdoor mobile-mapping data, and no systematic comparisons. The present paper tries to close that gap. It reviews the building blocks needed to assemble a panoptic segmentation pipeline and the related literature. Moreover, a modular pipeline is set up to perform comprehensive, systematic experiments to assess the state of panoptic segmentation in the context of street mapping. As a byproduct, we also provide the first public dataset for that task, by extending the NPM3D dataset to include instance labels. That dataset and our source code are publicly available.1We discuss which adaptations are need to adapt current panoptic segmentation methods to outdoor scenes and large objects. Our study finds that for mobile mapping data, KPConv performs best but is slower, while PointNet++ is fastest but performs significantly worse. Sparse CNNs are in between. Regardless of the backbone, instance segmentation by clustering embedding features is better than using shifted coordinates.
引用
收藏
页码:373 / 391
页数:19
相关论文
共 103 条
  • [11] Hierarchical Aggregation for 3D Instance Segmentation
    Chen, Shaoyu
    Fang, Jiemin
    Zhang, Qian
    Liu, Wenyu
    Wang, Xinggang
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 15447 - 15456
  • [12] 4D Spatio-Temporal ConvNets: Minkowski Convolutional Neural Networks
    Choy, Christopher
    Gwak, JunYoung
    Savarese, Silvio
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 3070 - 3079
  • [13] ICM-3D: Instantiated Category Modeling for 3D Instance Segmentation
    Chu, Ruihang
    Chen, Yukang
    Kong, Tao
    Qi, Lu
    Li, Lei
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (01) : 57 - 64
  • [14] Mean shift: A robust approach toward feature space analysis
    Comaniciu, D
    Meer, P
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2002, 24 (05) : 603 - 619
  • [15] Cortinhal T, 2020, Arxiv, DOI arXiv:2003.03653
  • [16] 3DMV: Joint 3D-Multi-view Prediction for 3D Semantic Scene Segmentation
    Dai, Angela
    Niessner, Matthias
    [J]. COMPUTER VISION - ECCV 2018, PT X, 2018, 11214 : 458 - 474
  • [17] Shape Completion using 3D-Encoder-Predictor CNNs and Shape Synthesis
    Dai, Angela
    Qi, Charles Ruizhongtai
    Niessner, Matthias
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 6545 - 6554
  • [18] ScanNet: Richly-annotated 3D Reconstructions of Indoor Scenes
    Dai, Angela
    Chang, Angel X.
    Savva, Manolis
    Halber, Maciej
    Funkhouser, Thomas
    Niessner, Matthias
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 2432 - 2443
  • [19] De Brabandere B, 2017, Arxiv, DOI [arXiv:1708.02551, 10.48550/arXiv.1708.02551]
  • [20] Deep Learning for LiDAR Point Cloud Classification in Remote Sensing
    Diab, Ahmed
    Kashef, Rasha
    Shaker, Ahmed
    [J]. SENSORS, 2022, 22 (20)