Continuous Control for Automated Lane Change Behavior Based on Deep Deterministic Policy Gradient Algorithm

被引:0
作者
Wang, Pin [1 ]
Li, Hanhan [1 ]
Chan, Ching-Yao [1 ]
机构
[1] Univ Calif Berkeley, Calif Path, 1357 South 46th St, Richmond, CA 94804 USA
来源
2019 30TH IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV19) | 2019年
关键词
Deep Deterministic Policy Gradient; Continuous Control; Lane Change; Autonomous Driving;
D O I
10.1109/ivs.2019.8813903
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Lane change is a challenging task which requires delicate actions to ensure safety and comfort. Some recent studies have attempted to solve the lane-change control problem with Reinforcement Learning (RL), yet the action is confined to discrete action space. To overcome this limitation, we formulate the lane change behavior with continuous action in a model-free dynamic driving environment based on Deep Deterministic Policy Gradient (DDPG). The reward function, which is critical for learning the optimal policy, is defined by control values, position deviation status, and maneuvering time to provide the RL agent informative signals. The RL agent is trained from scratch without resorting to any prior knowledge of the environment and vehicle dynamics since they are not easy to obtain. Seven models under different hyperparameter settings are compared. A video showing the learning progress of the driving behavior is available(2). It demonstrates the RL vehicle agent initially runs out of road boundary frequently, but eventually has managed to smoothly and stably change to the target lane with a success rate of 100% under diverse driving situations in simulation.
引用
收藏
页码:1454 / 1460
页数:7
相关论文
共 17 条
[1]  
Hoel C., 2018, 21 INT C INT TRANSP
[2]  
Jaritz M, 2018, IEEE INT CONF ROBOT, P2070
[3]  
Kaushik M., 2018, 29 IEEE INT VEH S IV
[4]   Reinforcement learning in robotics: A survey [J].
Kober, Jens ;
Bagnell, J. Andrew ;
Peters, Jan .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2013, 32 (11) :1238-1274
[5]  
Levine Sergey, 2013, INT C MACH LEARN, V28, P1
[6]  
Liang X., 2018, P EUR C COMP VIS, P584
[7]  
Lillicrap T. P., 2016, 6 INT C LEARN REPR I
[8]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533
[9]  
Mukadam M., 2017, NIPS WORKSH MACH LEA, P1
[10]  
Riedmiller M., 2018, LEARNING PLAYING SOL