End-to-End Autonomous Driving With Semantic Depth Cloud Mapping and Multi-Agent

被引:25
作者
Natan, Oskar [1 ,2 ]
Miura, Jun [1 ]
机构
[1] Toyohashi Univ Technol, Dept Comp Sci & Engn, Toyohashi 4418580, Japan
[2] Gadjah Mada Univ, Dept Comp Sci & Elect, Yogyakarta 55281, Indonesia
来源
IEEE TRANSACTIONS ON INTELLIGENT VEHICLES | 2023年 / 8卷 / 01期
关键词
Semantics; Task analysis; Feature extraction; Training; Multitasking; Sensors; Computational modeling; End-to-end deep learning; imitation learning; semantic depth cloud; multi-agent; autonomous driving; DEEP NEURAL-NETWORK;
D O I
10.1109/TIV.2022.3185303
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Focusing on the task of point-to-point navigation for an autonomous driving vehicle, we propose a novel deep learning model trained with end-to-end and multi-task learning manners to perform both perception and control tasks simultaneously. The model is used to drive the ego vehicle safely by following a sequence of routes defined by the global planner. The perception part of the model is used to encode high-dimensional observation data provided by an RGBD camera while performing semantic segmentation, semantic depth cloud (SDC) mapping, and traffic light state and stop sign prediction. Then, the control part decodes the encoded features along with additional information provided by GPS and speedometer to predict waypoints that come with a latent feature space. Furthermore, two agents are employed to process these outputs and make a control policy that determines the level of steering, throttle, and brake as the final action. The model is evaluated on CARLA simulator with various scenarios made of normal-adversarial situations and different weathers to mimic real-world conditions. In addition, we do a comparative study with some recent models to justify the performance in multiple aspects of driving. Moreover, we also conduct an ablation study on SDC mapping and multi-agent to understand their roles and behavior. As a result, our model achieves the highest driving score even with fewer parameters and computation load. To support future studies, we share our codes at https://github.com/oskarnatan/end-to-end-driving.
引用
收藏
页码:557 / 571
页数:15
相关论文
共 57 条
[1]  
Ayache N, 2017, 2017 INTELLIGENT SYSTEMS AND COMPUTER VISION (ISCV)
[2]  
Best A, 2017, IEEE INT C INT ROBOT, P2629, DOI 10.1109/IROS.2017.8206087
[3]  
Bicer Y, 2019, IEEE INT C INT ROBOT, P2629, DOI [10.1109/iros40897.2019.8967948, 10.1109/IROS40897.2019.8967948]
[4]  
Cao J., 2020, PROC INTERCONF UK CH, P1
[5]  
Chen D., 2019, PROC ANN C ROBOT LEA, P1
[6]   Interpretable End-to-End Urban Autonomous Driving With Latent Deep Reinforcement Learning [J].
Chen, Jianyu ;
Li, Shengbo Eben ;
Tomizuka, Masayoshi .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (06) :5068-5078
[7]   Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation [J].
Chen, Liang-Chieh ;
Zhu, Yukun ;
Papandreou, George ;
Schroff, Florian ;
Adam, Hartwig .
COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 :833-851
[8]   Stabilization Approaches for Reinforcement Learning-Based End-to-End Autonomous Driving [J].
Chen, Siyuan ;
Wang, Meiling ;
Song, Wenjie ;
Yang, Yi ;
Li, Yujun ;
Fu, Mengyin .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (05) :4740-4750
[9]  
Chen ZL, 2017, IEEE INT VEH SYM, P1856, DOI 10.1109/IVS.2017.7995975
[10]  
Cho K., 2014, C EMP METH NAT LANG, DOI DOI 10.3115/V1/W14-4012