A Review of panoptic segmentation for mobile mapping point clouds

被引:4
作者
Xiang, Binbin [1 ]
Yue, Yuanwen [1 ]
Peters, Torben [1 ]
Schindler, Konrad [1 ]
机构
[1] Swiss Fed Inst Technol, Photogrammetry & Remote Sensing, CH-8093 Zurich, Switzerland
关键词
Mobile mapping point clouds; 3D panoptic segmentation; 3D semantic segmentation; 3D instance segmentation; 3D deep learning backbones; CLASSIFICATION; NETWORKS; DATASET;
D O I
10.1016/j.isprsjprs.2023.08.008
中图分类号
P9 [自然地理学];
学科分类号
0705 ; 070501 ;
摘要
3D point cloud panoptic segmentation is the combined task to (i) assign each point to a semantic class and (ii) separate the points in each class into object instances. Recently there has been an increased interest in such comprehensive 3D scene understanding, building on the rapid advances of semantic segmentation due to the advent of deep 3D neural networks. Yet, to date there is very little work about panoptic segmentation of outdoor mobile-mapping data, and no systematic comparisons. The present paper tries to close that gap. It reviews the building blocks needed to assemble a panoptic segmentation pipeline and the related literature. Moreover, a modular pipeline is set up to perform comprehensive, systematic experiments to assess the state of panoptic segmentation in the context of street mapping. As a byproduct, we also provide the first public dataset for that task, by extending the NPM3D dataset to include instance labels. That dataset and our source code are publicly available.1We discuss which adaptations are need to adapt current panoptic segmentation methods to outdoor scenes and large objects. Our study finds that for mobile mapping data, KPConv performs best but is slower, while PointNet++ is fastest but performs significantly worse. Sparse CNNs are in between. Regardless of the backbone, instance segmentation by clustering embedding features is better than using shifted coordinates.
引用
收藏
页码:373 / 391
页数:19
相关论文
共 103 条
  • [1] A Survey on Deep-Learning-Based LiDAR 3D Object Detection for Autonomous Driving
    Alaba, Simegnew Yihunie
    Ball, John E.
    [J]. SENSORS, 2022, 22 (24)
  • [2] 3D Semantic Parsing of Large-Scale Indoor Spaces
    Armeni, Iro
    Sener, Ozan
    Zamir, Amir R.
    Jiang, Helen
    Brilakis, Ioannis
    Fischer, Martin
    Savarese, Silvio
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 1534 - 1543
  • [3] Towards 3D LiDAR-based semantic scene understanding of 3D point cloud sequences: The SemanticKITTI Dataset
    Behley, Jens
    Garbade, Martin
    Milioto, Andres
    Quenzel, Jan
    Behnke, Sven
    Gall, Juergen
    Stachniss, Cyrill
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2021, 40 (8-9) : 959 - 967
  • [4] SemanticKITTI: A Dataset for Semantic Scene Understanding of LiDAR Sequences
    Behley, Jens
    Garbade, Martin
    Milioto, Andres
    Quenzel, Jan
    Behnke, Sven
    Stachniss, Cyrill
    Gall, Juergen
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 9296 - 9306
  • [5] Review: Deep Learning on 3D Point Clouds
    Bello, Saifullahi Aminu
    Yu, Shangshu
    Wang, Cheng
    Adam, Jibril Muhmmad
    Li, Jonathan
    [J]. REMOTE SENSING, 2020, 12 (11)
  • [6] Pointwise Convolutional Neural Networks
    Binh-Son Hua
    Minh-Khoi Tran
    Yeung, Sai-Kit
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 984 - 993
  • [7] Burume D., 2021, Deep learning methods applied to 3D point clouds based instance segmentation: A review, DOI DOI 10.20944/PREPRINTS202111.0228.V1
  • [8] Caesar H, 2020, PROC CVPR IEEE, P11618, DOI 10.1109/CVPR42600.2020.01164
  • [9] Campello Ricardo J. G. B., 2013, Advances in Knowledge Discovery and Data Mining. 17th Pacific-Asia Conference (PAKDD 2013). Proceedings, P160, DOI 10.1007/978-3-642-37456-2_14
  • [10] Torch-Points3D: A Modular Multi-Task Framework for Reproducible Deep Learning on 3D Point Clouds
    Chaton, Thomas
    Chaulet, Nicolas
    Horache, Sofiane
    Landrieu, Loic
    [J]. 2020 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2020), 2020, : 190 - 199