Decision-Making for Autonomous Vehicles in Random Task Scenarios at Unsignalized Intersection Using Deep Reinforcement Learning

被引:5
作者
Xiao, Wenxuan [1 ]
Yang, Yuyou [1 ]
Mu, Xinyu [1 ]
Xie, Yi [1 ]
Tang, Xiaolin [1 ]
Cao, Dongpu [2 ]
Liu, Teng [3 ,4 ,5 ]
机构
[1] Chongqing Univ, Coll Mech & Vehicle Engn, Chongqing 400044, Peoples R China
[2] Univ Waterloo, Dept Mech & Mechatron Engn, Waterloo, ON N2L 3G1, Canada
[3] Chongqing Univ, Clin Res Ctr CRC, Clin Pathol Ctr CPC, Canc Early Detect Treatment Ctr, Chongqing 404000, Peoples R China
[4] Chongqing Univ, Gorges Hosp 3, Translat Med Res Ctr TMRC, Chongqing 404000, Peoples R China
[5] Chongqing Univ, Coll Mech & Vehicle Engn, Chongqing 400044, Peoples R China
关键词
Task analysis; Decision making; Autonomous vehicles; Behavioral sciences; Reinforcement learning; Turning; Safety; Deep reinforcement learning; random driving task; decision-making; autonomous vehicles; unsignalized intersection;
D O I
10.1109/TVT.2024.3360445
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This study constructs a decision-making control framework for autonomous ego vehicles (AEVs) based on Soft Actor-Critic (SAC) in a random driving task scenario at an unsignalized intersection. The environment vehicles include both AEV and surrounding vehicles, and the three driving tasks through unsignalized intersections are going straight, left turn, and right turn. Since the driving tasks of AEV and surrounding vehicles are random, the environment is characterized by high uncertainty and difficulty. There are three innovative points in this paper. First, this paper proposes a new Mix-Attention Network based on the attention mechanism. Second, this paper improves the state by introducing a new input quantity to represent the driving task of the vehicle itself. Third, this paper has been enhanced in replay buffer, using more collision and arrival experiences to train the neural network. In this paper, the performance of the original and improved models is evaluated in terms of safety and efficiency. The simulation results show that all three proposed improvement methods can improve performance and achieve better results.
引用
收藏
页码:7812 / 7825
页数:14
相关论文
共 32 条
[1]   Survey of Deep Reinforcement Learning for Motion Planning of Autonomous Vehicles [J].
Aradi, Szilard .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (02) :740-759
[2]  
Bouton M., 2019, arXiv
[3]  
Bouton M, 2017, IEEE INT VEH SYM, P825, DOI 10.1109/IVS.2017.7995818
[4]  
Chen C, 2016, IEEE INT VEH SYM, P1242, DOI 10.1109/IVS.2016.7535549
[5]   Cooperative Intersection Management: A Survey [J].
Chen, Lei ;
Englund, Cristofer .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2016, 17 (02) :570-586
[6]  
Christodoulou P, 2019, Arxiv, DOI arXiv:1910.07207
[7]   Proximal Policy Optimization Through a Deep Reinforcement Learning Framework for Multiple Autonomous Vehicles at a Non-Signalized Intersection [J].
Duy Quang Tran ;
Bae, Sang-Hoon .
APPLIED SCIENCES-BASEL, 2020, 10 (16)
[8]   Automated Intersection Control Performance of Future Innovation Versus Current Traffic Signal Control [J].
Fajardo, David ;
Au, Tsz-Chiu ;
Waller, S. Travis ;
Stone, Peter ;
Yang, David .
TRANSPORTATION RESEARCH RECORD, 2011, (2259) :223-232
[9]  
Haarnoja T, 2019, Arxiv, DOI arXiv:1812.05905
[10]  
Haarnoja T, 2018, PR MACH LEARN RES, V80