Obstacle avoidance planning of autonomous vehicles using deep reinforcement learning

被引:6
作者
Qian, Yubin [1 ]
Feng, Song [1 ]
Hu, Wenhao [2 ]
Wang, Wanqiu [1 ]
机构
[1] Shanghai Univ Engn Sci, Sch Mech & Automot Engn, Songjiang Campus LongTeng Rd 333, Shanghai 201620, Peoples R China
[2] Defect Prod Adm Ctr SAMR, Beijing, Peoples R China
关键词
Obstacle avoidance; autonomous vehicle; path planning; deep reinforcement learning; long and short-term memory networks; HIGHWAY; RRT;
D O I
10.1177/16878132221139661
中图分类号
O414.1 [热力学];
学科分类号
摘要
Obstacle avoidance path planning in a dynamic circumstance is one of the fundamental problems of autonomous vehicles, counting optional maneuvers: emergency braking and active steering. This paper proposes emergency obstacle avoidance planning based on deep reinforcement learning (DRL), considering safety and comfort. Firstly, the vehicle emergency braking and lane change processes are analyzed in detail. A graded hazard index is defined to indicate the degree of the potential risk of the current vehicle movement. The longitudinal distance and lateral waypoint models are established, including the comfort deceleration and stability coefficient considerations. Simultaneously, a fuzzy PID controller is installed to track to satisfy the stability and feasibility of the path. Then, this paper proposes a DRL process to determine the obstacle avoidance plan. Mainly, multi-reward functions are designed for different collisions, corresponding penalties for longitudinal rear-end collisions, and lane-changing side collisions based on the safety distance, comfort reward, and safety reward. Apply a special DRL method-DQN to release the planning program. The difference is that the long and short-term memory (LSTM) layer is utilized to solve incomplete observations and improve the efficiency and stability of the algorithm in a dynamic environment. Once the policy is practiced, the vehicle can automatically perform the best obstacle avoidance maneuver in an emergency, improving driving safety. Finally, this paper builds a simulated environment in CARLA and is trained to evaluate the effectiveness of the proposed algorithm. The collision rate, safety distance difference, and total reward index indicate that the collision avoidance path is generated safely, and the lateral acceleration and longitudinal velocity satisfy the comfort requirements. Besides, the method proposed in this paper is compared with traditional DRL, which proves the beneficial performance in safety and efficiency.
引用
收藏
页数:14
相关论文
共 33 条
[1]   Trajectory Planning for Autonomous Vehicles Using Hierarchical Reinforcement Learning [J].
Ben Naveed, Kaleb ;
Qiao, Zhiqian ;
Dolan, John M. .
2021 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2021, :601-606
[2]   Adaptive speed control method for electromagnetic direct drive vehicle robot driver based on fuzzy logic [J].
Chen, Gang ;
Zhang, Weigong ;
Li, Xu ;
Yu, Bing .
MEASUREMENT & CONTROL, 2019, 52 (9-10) :1344-1353
[3]   An improved A-Star based path planning algorithm for autonomous land vehicles [J].
Erke, Shang ;
Bin, Dai ;
Yiming, Nie ;
Qi, Zhu ;
Liang, Xiao ;
Dawei, Zhao .
INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2020, 17 (05)
[4]   A Decision-Making Strategy for Vehicle Autonomous Braking in Emergency via Deep Reinforcement Learning [J].
Fu, Yuchuan ;
Li, Changle ;
Yu, Fei Richard ;
Luan, Tom H. ;
Zhang, Yao .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (06) :5876-5888
[5]   Evaluating semi-cooperative Nash/Stackelberg Q-learning for traffic routes plan in a single intersection [J].
Guo, Jian ;
Harmati, Istvan .
CONTROL ENGINEERING PRACTICE, 2020, 102
[6]  
Hochreiter S, 1997, NEURAL COMPUT, V9, P1735, DOI [10.1162/neco.1997.9.8.1735, 10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]
[7]   Combining Planning and Deep Reinforcement Learning in Tactical Decision Making for Autonomous Driving [J].
Hoel, Carl-Johan ;
Driggs-Campbell, Katherine ;
Wolff, Krister ;
Laine, Leo ;
Kochenderfer, Mykel J. .
IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2020, 5 (02) :294-305
[8]   An Improved Artificial Potential Field Model Considering Vehicle Velocity for Autonomous Driving [J].
Hu Hongyu ;
Zhang Chi ;
Sheng Yuhuan ;
Zhou Bin ;
Gao Fei .
IFAC PAPERSONLINE, 2018, 51 (31) :863-867
[9]   Convolutional neural networkscheme-based optical camera communication system for intelligent Internet of vehicles [J].
Islam, Amirul ;
Hossan, Md Tanvir ;
Jang, Yeong Min .
INTERNATIONAL JOURNAL OF DISTRIBUTED SENSOR NETWORKS, 2018, 14 (04)
[10]   Path Planning and Tracking for Vehicle Collision Avoidance Based on Model Predictive Control With Multiconstraints [J].
Ji, Jie ;
Khajepour, Amir ;
Melek, Wael William ;
Huang, Yanjun .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2017, 66 (02) :952-964