ES-DQN: A Learning Method for Vehicle Intelligent Speed Control Strategy Under Uncertain Cut-In Scenario

被引:24
作者
Chen, Qingyun [1 ]
Zhao, Wanzhong [1 ]
Li, Lin [1 ]
Wang, Chunyan [1 ]
Chen, Feng [1 ,2 ]
机构
[1] Nanjing Univ Aeronaut & Astronaut, Vehicle Engn, Nanjing 210016, Peoples R China
[2] Zhejiang Vie Sci & Technol Co Ltd, Zhuji City 311835, Zhejiang, Peoples R China
关键词
Velocity control; Autonomous vehicles; Control systems; Q-learning; Training; Learning systems; Wheels; Intelligent speed control strategy; deep Q-learning network; experience screening; reinforcement learning; cut-in scenarios;
D O I
10.1109/TVT.2022.3143840
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Uncertain cut-in maneuver of vehicles from adjacent lanes makes it difficult for vehicle's automatic speed control strategy to make judgments and effective control decisions. In this paper, an intelligent speed control strategy for uncertain cut-in scenarios is established based on a basic autonomous driving system. This strategy judges cut-in maneuver from surrounding vehicles and outputs adaptive control action under current environment according to Q value of state-action pair based on a Q network. In addition, according to the analysis of cut-in scenarios, the Q network is trained based on a novel reinforcement learning method named as experience screening deep Q-learning network (ES-DQN). The proposed ES-DQN is an extension of double deep Q-learning network (DDQN) algorithm, and includes two parts: experience screening and policy learning. Based on the experience screened from the experience screening part, the proposed learning method can train an intelligent speed control strategy which has stronger adaptability and control effect in uncertain cut-in scenarios. According to simulation results, the proposed intelligent speed control strategy trained by ES-DQN has better performance under uncertain cut-in scenarios than DDQN method and traditional ACC strategy. Meanwhile, by adjusting weight value in reward function, the system can realize different control target.
引用
收藏
页码:2472 / 2484
页数:13
相关论文
共 25 条
[1]   Scenario-based stochastic MPC for vehicle speed control considering the interaction with pedestrians [J].
Anh-Tuan Tran ;
Muraleedharan, Arun ;
Okuda, Hiroyuki ;
Suzuki, Tatsuya .
IFAC PAPERSONLINE, 2020, 53 (02) :15325-15331
[2]   Motion Planning With Velocity Prediction and Composite Nonlinear Feedback Tracking Control for Lane-Change Strategy of Autonomous Vehicles [J].
Chen, Yimin ;
Hu, Chuan ;
Wang, Junmin .
IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2020, 5 (01) :63-74
[3]   Cooperative Adaptive Cruise Control: A Reinforcement Learning Approach [J].
Desjardins, Charles ;
Chaib-draa, Brahim .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2011, 12 (04) :1248-1260
[4]  
Ding N., 2015, IEEE T IND INFORM, V14, P1
[5]   Reinforcement-Learning-Based Cooperative Adaptive Cruise Control of Buses in the Lincoln Tunnel Corridor with Time-Varying Topology [J].
Gao, Weinan ;
Gao, Jingqin ;
Ozbay, Kaan ;
Jiang, Zhong-Ping .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2019, 20 (10) :3796-3805
[6]  
Hu C., 2020, PROC I MECH ENG PART, V235, P1
[7]   Q-Learning Algorithms: A Comprehensive Classification and Applications [J].
Jang, Beakcheol ;
Kim, Myeonghwi ;
Harerimana, Gaspard ;
Kim, Jong Wook .
IEEE ACCESS, 2019, 7 :133653-133667
[8]   Policy gradient reinforcement learning for fast quadrupedal locomotion [J].
Kohl, N ;
Stone, P .
2004 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1- 5, PROCEEDINGS, 2004, :2619-2624
[9]   Ecological Adaptive Cruise Control for Vehicles With Step-Gear Transmission Based on Reinforcement Learning [J].
Li, Guoqiang ;
Gorges, Daniel .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2020, 21 (11) :4895-4905
[10]   Training Drift Counteraction Optimal Control Policies Using Reinforcement Learning: An Adaptive Cruise Control Example [J].
Li, Zhaojian ;
Chu, Tianshu ;
Kolmanovsky, Ilya, V ;
Yin, Xiang .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2018, 19 (09) :2903-2912