Highway Exiting Planner for Automated Vehicles Using Reinforcement Learning

被引:48
作者
Cao, Zhong [1 ,2 ]
Yang, Diange [1 ]
Xu, Shaobing [2 ]
Peng, Huei [2 ]
Li, Boqi [2 ]
Feng, Shuo [1 ]
Zhao, Ding [3 ]
机构
[1] Tsinghua Univ, Dept Automot Engn, Beijing 100084, Peoples R China
[2] Univ Michigan, Dept Mech Engn, Ann Arbor, MI 48105 USA
[3] Carnegie Mellon Univ, Dept Mech Engn, Pittsburgh, PA 15213 USA
关键词
Road transportation; Safety; Reinforcement learning; Vehicle dynamics; Vehicles; Dynamics; Trajectory; Autonomous vehicle; motion planning; decision making; reinforcement learning; LANE; MODEL;
D O I
10.1109/TITS.2019.2961739
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Exiting from highways in crowded dynamic traffic is an important path planning task for autonomous vehicles (AVs). This task can be challenging because of the uncertain motion of surrounding vehicles and limited sensing/observing window. Conventional path planning methods usually compute a mandatory lane change (MLC) command, but the lane change behavior (e.g., vehicle speed and gap acceptance) should also adapt to traffic conditions and the urgency for exiting. In this paper, we propose a reinforcement learning-enhanced highway-exit planner. The learning-based strategy learns from past failures and adjusts the vehicle motion when the AV fails to exit. The reinforcement learning is based on the Monte Carlo tree search (MCTS) approach. The proposed learning-enhanced highway-exit planner is tested 6000 times in stochastic simulations. The results indicate that the proposed planner achieves a higher probability of successful highway exiting than a benchmark MLC planner.
引用
收藏
页码:990 / 1000
页数:11
相关论文
共 32 条
[11]   Modeling Mandatory Lane Changing Using Bayes Classifier and Decision Trees [J].
Hou, Yi ;
Edara, Praveen ;
Sun, Carlos .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2014, 15 (02) :647-655
[12]   General lane-changing model MOBIL for car-following models [J].
Kesting, Arne ;
Treiber, Martin ;
Helbing, Dirk .
TRANSPORTATION RESEARCH RECORD, 2007, (1999) :86-94
[13]  
Koutnik J., 2014, P INT C SIM AD BEH
[14]   Training Drift Counteraction Optimal Control Policies Using Reinforcement Learning: An Adaptive Cruise Control Example [J].
Li, Zhaojian ;
Chu, Tianshu ;
Kolmanovsky, Ilya, V ;
Yin, Xiang .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2018, 19 (09) :2903-2912
[15]   Optimal adaptive cruise control with guaranteed string stability [J].
Liang, CY ;
Peng, H .
VEHICLE SYSTEM DYNAMICS, 1999, 32 (4-5) :313-330
[16]  
Liu CR, 2017, IEEE INT C INTELL TR
[17]   Lane Change Maneuvers for Automated Vehicles [J].
Nilsson, Julia ;
Brannstrom, Mattias ;
Coelingh, Erik ;
Fredriksson, Jonas .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2017, 18 (05) :1087-1096
[18]   If, When, and How to Perform Lane Change Maneuvers on Highways [J].
Nilsson, Julia ;
Silvlin, Jonatan ;
Brannstrom, Mattias ;
Coelingh, Erik ;
Fredriksson, Jonas .
IEEE INTELLIGENT TRANSPORTATION SYSTEMS MAGAZINE, 2016, 8 (04) :68-78
[19]  
Nilsson J, 2015, P AMER CONTR CONF, P1399, DOI 10.1109/ACC.2015.7170929
[20]   Impact of on-board tutoring systems to improve driving efficiency of non-professional drivers [J].
Pozueco, Laura ;
Rionda, Abel ;
Paneda, Alejandro Garca ;
Sanchez, Jose Antonio ;
Paneda, Xabiel Garcia ;
Garcia, Roberto ;
Melendi, David ;
Tuero, Alejandro Garcia .
IET INTELLIGENT TRANSPORT SYSTEMS, 2017, 11 (04) :196-202