Rate GQN: A Deviations-Reduced Decision-Making Strategy for Connected and Automated Vehicles in Mixed Autonomy

被引:9
作者
Gao, Xin [1 ]
Li, Xueyuan [1 ]
Liu, Qi [1 ]
Ma, Zhaoyang [2 ]
Luan, Tian [1 ]
Yang, Fan [1 ]
Li, Zirui [1 ,3 ]
机构
[1] Beijing Inst Technol, Sch Mech Engn, Beijing 100081, Peoples R China
[2] Beijing Jiaotong Univ, Sch Comp & Informat Technol, Beijing 100044, Peoples R China
[3] Tech Univ Dresden, FriedrichList Fac Transport & Traff Sci, Chair Traff Proc Automat, D-01062 Dresden, Germany
关键词
Rate graph convolution Q-learning network; connected autonomous vehicles; internal dynamic multi-objective reward function; spatial-temporal interaction; DRIVER; ROAD;
D O I
10.1109/TITS.2023.3312951
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Connected and automated vehicles (CAVs) have become one of the essential approaches to effectively resolve problems such as traffic safety, road congestion, and energy consumption. However, due to the spatial-temporal interaction of the mixed traffic environment, the driving behaviors of traffic participants are continuously transmitted in time and space. This makes it difficult for the existing decision-making system of CAVs to make accurate judgments and effective strategies. In this study, a rate graph convolution Q-learning network (Rate GQN) is proposed to train a discrete strategy that can improve the comprehensive performance of CAVs in scenarios with spatial-temporal interaction. Firstly, the Rate algorithm is proposed to impose a ratio on the estimates of Q-values from the previous learning process, which improves the stability and performance of the algorithm by reducing the approximate error variance of the target value. Secondly, the traffic Scenario is modeled as a graph structure. And graph convolutional networks are adopted to extract the features information of graph structure to help the CAVs grasp the dynamic traffic interaction information quickly and accurately. Additionally, an internal dynamic multi-objective reward function is presented to improve the comprehensive performance of CAVs, including safety, efficiency, energy saving, and comfort. Finally, comparison and ablation experiments are constructed in a task-based traffic scenario (station stop and traffic light passing). The simulation results show that our Rate GQN method has faster training speed, a more stable training process, and better overall performance than the deep Q-learning network (DQN) and algorithms of the comparison group.
引用
收藏
页码:613 / 625
页数:13
相关论文
共 42 条
[1]  
Anschel O, 2017, 34 INT C MACHINE LEA, V70
[2]   Multi-Agent Deep Reinforcement Learning to Manage Connected Autonomous Vehicles at Tomorrow's Intersections [J].
Antonio, Guillen-Perez ;
Maria-Dolores, Cano .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (07) :7033-7043
[3]   Hybrid Reinforcement Learning-Based Eco-Driving Strategy for Connected and Automated Vehicles at Signalized Intersections [J].
Bai, Zhengwei ;
Hao, Peng ;
Shangguan, Wei ;
Cai, Baigen ;
Barth, Matthew J. .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (09) :15850-15863
[4]  
Bouton M, 2019, IEEE INT C INTELL TR, P3441, DOI [10.1109/ITSC.2019.8916924, 10.1109/itsc.2019.8916924]
[5]   The Planning Process of Transport Tasks for Autonomous Vans-Case Study [J].
Caban, Jacek ;
Nieoczym, Aleksander ;
Dudziak, Agnieszka ;
Krajka, Tomasz ;
Stopkova, Maria .
APPLIED SCIENCES-BASEL, 2022, 12 (06)
[6]   Past, Present, and Future of Simultaneous Localization and Mapping: Toward the Robust-Perception Age [J].
Cadena, Cesar ;
Carlone, Luca ;
Carrillo, Henry ;
Latif, Yasir ;
Scaramuzza, Davide ;
Neira, Jose ;
Reid, Ian ;
Leonard, John J. .
IEEE TRANSACTIONS ON ROBOTICS, 2016, 32 (06) :1309-1332
[7]   Highway Exiting Planner for Automated Vehicles Using Reinforcement Learning [J].
Cao, Zhong ;
Yang, Diange ;
Xu, Shaobing ;
Peng, Huei ;
Li, Boqi ;
Feng, Shuo ;
Zhao, Ding .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (02) :990-1000
[8]   Joint Optimization of Sensing, Decision-Making and Motion-Controlling for Autonomous Vehicles: A Deep Reinforcement Learning Approach [J].
Chen, Longquan ;
He, Ying ;
Wang, Qiang ;
Pan, Weike ;
Ming, Zhong .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (05) :4642-4654
[9]   ES-DQN: A Learning Method for Vehicle Intelligent Speed Control Strategy Under Uncertain Cut-In Scenario [J].
Chen, Qingyun ;
Zhao, Wanzhong ;
Li, Lin ;
Wang, Chunyan ;
Chen, Feng .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (03) :2472-2484
[10]  
Dong JQ, 2020, Arxiv, DOI arXiv:2010.05437