Multi-Class Road User Detection With 3+1D Radar in the View-of-Delft Dataset

被引:135
作者
Palffy, Andras [1 ]
Pool, Ewoud [1 ]
Baratam, Srimannarayana [1 ]
Kooij, Julian F. P. [1 ]
Gavrila, Dariu M. [1 ]
机构
[1] Delft Univ Technol, Intelligent Vehicles Grp, NL-2628 CD Delft, Netherlands
关键词
Object detection; segmentation and categorization; data sets for robotic vision; automotive radars; AUTOMOTIVE RADAR; CAMERA;
D O I
10.1109/LRA.2022.3147324
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Next-generation automotive radars provide elevation data in addition to range-, azimuth- and Doppler velocity. In this experimental study, we apply a state-of-the-art object detector (PointPillars), previously used for LiDAR 3D data, to such 3+1D radar data (where 1D refers to Doppler). In ablation studies, we first explore the benefits of the additional elevation information, together with that of Doppler, radar cross section and temporal accumulation, in the context of multi-class road user detection. We subsequently compare object detection performance on the radar and LiDAR point clouds, object class-wise and as a function of distance. To facilitate our experimental study, we present the novel View-of-Delft (VoD) automotive dataset. It contains 8693 frames of synchronized and calibrated 64-layer LiDAR-, (stereo) camera-, and 3+1D radar-data acquired in complex, urban traffic. It consists of 123106 3D bounding box annotations of both moving and static objects, including 26587 pedestrian, 10800 cyclist and 26949 car labels. Our results show that object detection on 64-layer LiDAR data still outperforms that on 3+1D radar data, but the addition of elevation information and integration of successive radar scans helps close the gap. The VoD dataset is made freely available for scientific benchmarking at https://intelligent-vehicles.org/datasets/view-of-delft/.
引用
收藏
页码:4961 / 4968
页数:8
相关论文
共 41 条
[11]  
Ferranti L, 2019, IEEE INT VEH SYM, P1660, DOI 10.1109/IVS.2019.8813899
[12]   Vision meets robotics: The KITTI dataset [J].
Geiger, A. ;
Lenz, P. ;
Stiller, C. ;
Urtasun, R. .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2013, 32 (11) :1231-1237
[13]  
Held P, 2019, 2019 IEEE INTERNATIONAL CONFERENCE OF VEHICULAR ELECTRONICS AND SAFETY (ICVES 19)
[14]   SO-Net: Joint Semantic Segmentation and Obstacle Detection Using Deep Fusion of Monocular Camera and Radar [J].
John, V ;
Nithilan, M. K. ;
Mita, S. ;
Tehrani, H. ;
Sudheesh, R. S. ;
Lalu, P. P. .
IMAGE AND VIDEO TECHNOLOGY, PSIVT 2019 INTERNATIONAL WORKSHOPS, 2020, 11994 :138-148
[15]   RVNet: Deep Sensor Fusion of Monocular Camera and Radar for Image-Based Obstacle Detection in Challenging Environments [J].
John, Vijay ;
Mita, Seiichi .
IMAGE AND VIDEO TECHNOLOGY (PSIVT 2019), 2019, 11854 :351-364
[16]  
Kellner D, 2013, 2013 16TH INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION), P877
[17]   MulRan: Multimodal Range Dataset for Urban Place Recognition [J].
Kim, Giseop ;
Park, Yeong Sang ;
Cho, Younghun ;
Jeong, Jinyong ;
Kim, Ayoung .
2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, :6246-6253
[18]   PointPillars: Fast Encoders for Object Detection from Point Clouds [J].
Lang, Alex H. ;
Vora, Sourabh ;
Caesar, Holger ;
Zhou, Lubing ;
Yang, Jiong ;
Beijbom, Oscar .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :12689-12697
[19]  
Lombacher J, 2016, 2016 IEEE MTT-S INTERNATIONAL CONFERENCE ON MICROWAVES FOR INTELLIGENT MOBILITY (ICMIM)
[20]  
Lombacher J, 2017, IEEE INT VEH SYM, P1170, DOI 10.1109/IVS.2017.7995871