Enhanced Perception for Autonomous Driving Using Semantic and Geometric Data Fusion

被引:13
作者
Florea, Horatiu [1 ]
Petrovai, Andra [1 ]
Giosan, Ion [1 ]
Oniga, Florin [1 ]
Varga, Robert [1 ]
Nedevschi, Sergiu [1 ]
机构
[1] Tech Univ Cluj Napoca, Image Proc & Pattern Recognit Res Ctr, Comp Sci Dept, Cluj Napoca 400114, Romania
关键词
autonomous driving; environment perception; low-level geometry and semantic fusion; semantic and instance segmentation; deep learning; 3D object detection; TRACKING;
D O I
10.3390/s22135061
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Environment perception remains one of the key tasks in autonomous driving for which solutions have yet to reach maturity. Multi-modal approaches benefit from the complementary physical properties specific to each sensor technology used, boosting overall performance. The added complexity brought on by data fusion processes is not trivial to solve, with design decisions heavily influencing the balance between quality and latency of the results. In this paper we present our novel real-time, 360 degrees enhanced perception component based on low-level fusion between geometry provided by the LiDAR-based 3D point clouds and semantic scene information obtained from multiple RGB cameras, of multiple types. This multi-modal, multi-sensor scheme enables better range coverage, improved detection and classification quality with increased robustness. Semantic, instance and panoptic segmentations of 2D data are computed using efficient deep-learning-based algorithms, while 3D point clouds are segmented using a fast, traditional voxel-based solution. Finally, the fusion obtained through point-to-image projection yields a semantically enhanced 3D point cloud that allows enhanced perception through 3D detection refinement and 3D object classification. The planning and control systems of the vehicle receives the individual sensors' perception together with the enhanced one, as well as the semantically enhanced 3D points. The developed perception solutions are successfully integrated onto an autonomous vehicle software stack, as part of the UP-Drive project.
引用
收藏
页数:22
相关论文
共 54 条
[1]  
[Anonymous], UP DRIVE H2020 EUROP
[2]  
[Anonymous], EB ASSIST ADTF AUTOM
[3]   A Fast and Log-Euclidean Polyaffine Framework for Locally Linear Registration [J].
Arsigny, Vincent ;
Commowick, Olivier ;
Ayache, Nicholas ;
Pennec, Xavier .
JOURNAL OF MATHEMATICAL IMAGING AND VISION, 2009, 33 (02) :222-238
[4]   3D Lidar-based static and moving obstacle detection in driving environments: An approach based on voxels and multi-region ground planes [J].
Asvadi, Alireza ;
Premebida, Cristiano ;
Peixoto, Paulo ;
Nunes, Urbano .
ROBOTICS AND AUTONOMOUS SYSTEMS, 2016, 83 :299-311
[5]  
Bieder F., 2020, ARXIV
[6]  
Bijelic Mario, 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Proceedings, P11679, DOI 10.1109/CVPR42600.2020.01170
[7]   YOLACT Real-time Instance Segmentation [J].
Bolya, Daniel ;
Zhou, Chong ;
Xiao, Fanyi ;
Lee, Yong Jae .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :9156-9165
[8]  
Bürki M, 2018, IEEE INT VEH SYM, P682, DOI 10.1109/IVS.2018.8500432
[9]  
Caesar H, 2020, PROC CVPR IEEE, P11618, DOI 10.1109/CVPR42600.2020.01164
[10]   Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation [J].
Chen, Liang-Chieh ;
Zhu, Yukun ;
Papandreou, George ;
Schroff, Florian ;
Adam, Hartwig .
COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 :833-851