Hierarchical framework integrating rapidly-exploring random tree with deep reinforcement learning for autonomous vehicle

被引:15
作者
Yu, Jiaxing [1 ,2 ]
Arab, Aliasghar [2 ]
Yi, Jingang [2 ]
Pei, Xiaofei [1 ]
Guo, Xuexun [1 ]
机构
[1] Wuhan Univ Technol, Hubei Key Lab Adv Technol Automot Components, Luogui Rd, Wuhan 430070, Hubei, Peoples R China
[2] Rutgers State Univ, Dept Mech & Aerosp Engn, 98 Brett Rd, Piscataway, NJ 08854 USA
基金
美国国家科学基金会;
关键词
Autonomous vehicle; Reinforcement learning; Rapidly-exploring random tree (RRT); Machine learning;
D O I
10.1007/s10489-022-04358-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper proposes a systematic driving framework where the decision making module of reinforcement learning (RL) is integrated with rapidly-exploring random tree (RRT) as motion planning. RL is used to generate local goals and semantic speed commands to control the longitudinal speed of a vehicle while rewards are designed for the driving safety and the traffic efficiency. Guaranteeing the driving comfort, RRT returns a feasible path to be followed by the vehicle with the speed commands. The scene decomposition approach is implemented to scale the deep neural network (DNN) to environments with multiple traffic participants and double deep Q-networks (DDQN) with prioritized experience replay (PER) is utilized to accelerate the training process. To handle the disturbance of the perception of the agent, we use an ensemble of neural networks to evaluate the uncertainty of decisions. It has shown that the proposed framework can tackle unexpected actions of traffic participants at an intersection yielding safe, comfort and efficient driving behaviors. Also, the ensemble of DDQN with PER is proved to be superior over standard DDQN in terms of learning efficiency and disturbance vulnerability.
引用
收藏
页码:16473 / 16486
页数:14
相关论文
共 33 条
[11]  
Hoel CJ, 2018, IEEE INT C INTELL TR, P2148, DOI 10.1109/ITSC.2018.8569568
[12]   A Dynamic Adjusting Reward Function Method for Deep Reinforcement Learning with Adjustable Parameters [J].
Hu, Zijian ;
Wan, Kaifang ;
Gao, Xiaoguang ;
Zhai, Yiwei .
MATHEMATICAL PROBLEMS IN ENGINEERING, 2019, 2019
[13]  
Isele D, 2018, IEEE INT CONF ROBOT, P2034
[14]  
Islam F., 2012, 2012 IEEE International Conference on Mechatronics and Automation (ICMA), P1651, DOI 10.1109/ICMA.2012.6284384
[15]  
Karaman S, 2011, IEEE INT CONF ROBOT, P1478
[16]  
Kochenderfer MJ, 2015, MIT LINCOLN LAB, P1
[17]  
Kurzer K, 2021, GEN DECISION MAKING
[18]   Real-Time Motion Planning With Applications to Autonomous Urban Driving [J].
Kuwata, Yoshiaki ;
Teo, Justin ;
Fiore, Gaston ;
Karaman, Sertac ;
Frazzoli, Emilio ;
How, Jonathan P. .
IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2009, 17 (05) :1105-1118
[19]   Development of a new integrated local trajectory planning and tracking control framework for autonomous ground vehicles [J].
Li, Xiaohui ;
Sun, Zhenping ;
Cao, Dongpu ;
Liu, Daxue ;
He, Hangen .
MECHANICAL SYSTEMS AND SIGNAL PROCESSING, 2017, 87 :118-137
[20]   Neural Network Approximation Based Near-Optimal Motion Planning With Kinodynamic Constraints Using RRT [J].
Li, Yang ;
Cui, Rongxin ;
Li, Zhijun ;
Xu, Demin .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2018, 65 (11) :8718-8729