Panoptic Nuscenes: A Large-Scale Benchmark for LiDAR Panoptic Segmentation and Tracking

被引:67
作者
Fong, Whye Kit [1 ]
Mohan, Rohit [2 ]
Hurtado, Juana Valeria [2 ]
Zhou, Lubing [1 ]
Caesar, Holger [1 ]
Beijbom, Oscar [1 ]
Valada, Abhinav [2 ]
机构
[1] Motional, Singapore 138633, Singapore
[2] Univ Freiburg, Dept Comp Sci, D-79085 Freiburg, Germany
关键词
Laser radar; Semantics; Task analysis; Annotations; Benchmark testing; Three-dimensional displays; Urban areas; List of keywords (from the RA Letters keyword list);
D O I
10.1109/LRA.2022.3148457
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Panoptic scene understanding and tracking of dynamic agents are essential for robots and automated vehicles to navigate in urban environments. As LiDARs provide accurate illumination-independent geometric depictions of the scene, performing these tasks using LiDAR point clouds provides reliable predictions. However, existing datasets lack diversity in the type of urban scenes and have a limited number of dynamic object instances which hinders both learning of these tasks as well as credible benchmarking of the developed methods. In this letter, we introduce the large-scale Panoptic nuScenes benchmark dataset that extends our popular nuScenes dataset with point-wise groundtruth annotations for semantic segmentation, panoptic segmentation, and panoptic tracking tasks. To facilitate comparison, we provide several strong baselines for each of these tasks on our proposed dataset. Moreover, we analyze the drawbacks of the existing metrics for panoptic tracking and propose the novel instance-centric PAT metric that addresses the concerns. We present exhaustive experiments that demonstrate the utility of Panoptic nuScenes compared to existing datasets and make the online evaluation server available at nuScenes.org. We believe that this extension will accelerate the research of novel methods for scene understanding of dynamic urban environments.
引用
收藏
页码:3795 / 3802
页数:8
相关论文
共 32 条
  • [1] 4D Panoptic LiDAR Segmentation
    Ayguen, Mehmet
    Osep, Aljosa
    Weber, Mark
    Maximov, Maxim
    Stachniss, Cyrill
    Behley, Jens
    Leal-Taixe, Laura
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 5523 - 5533
  • [2] Behley J., 2021, P INT C ROB AUT, P959
  • [3] SemanticKITTI: A Dataset for Semantic Scene Understanding of LiDAR Sequences
    Behley, Jens
    Garbade, Martin
    Milioto, Andres
    Quenzel, Jan
    Behnke, Sven
    Stachniss, Cyrill
    Gall, Juergen
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 9296 - 9306
  • [4] Caesar H, 2020, PROC CVPR IEEE, P11618, DOI 10.1109/CVPR42600.2020.01164
  • [5] Cattaneo D., 2021, ARXIV210305056
  • [6] Argoverse: 3D Tracking and Forecasting with Rich Maps
    Chang, Ming-Fang
    Lambert, John
    Sangkloy, Patsorn
    Singh, Jagjeet
    Bak, Slawomir
    Hartnett, Andrew
    Wang, De
    Carr, Peter
    Lucey, Simon
    Ramanan, Deva
    Hays, James
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 8740 - 8749
  • [7] Cheng R., 2021, P IEEE C COMP VIS PA
  • [8] Panoster: End-to-End Panoptic Segmentation of LiDAR Point Clouds
    Gasperini, Stefano
    Mahani, Mohammad-Ali Nikouei
    Marcos-Ramiro, Alvaro
    Navab, Nassir
    Tombari, Federico
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) : 3216 - 3223
  • [9] Vision meets robotics: The KITTI dataset
    Geiger, A.
    Lenz, P.
    Stiller, C.
    Urtasun, R.
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2013, 32 (11) : 1231 - 1237
  • [10] Geyer J., 2020, ARXIV