Enhancing Car-Following Performance in Traffic Oscillations Using Expert Demonstration Reinforcement Learning

被引:3
作者
Li, Meng [1 ,2 ]
Li, Zhibin [1 ]
Cao, Zehong [3 ]
机构
[1] Southeast Univ, Sch Transportat, Nanjing 210096, Peoples R China
[2] Nanyang Technol Univ, Sch Mech & Aerosp Engn, Singapore 639798, Singapore
[3] Univ South Australia, STEM, Adelaide, SA 5095, Australia
基金
中国国家自然科学基金;
关键词
Training; Oscillators; Trajectory; Task analysis; Cloning; Safety; Databases; Expert demonstration; reinforcement learning; car-following control; traffic oscillation; ADAPTIVE CRUISE CONTROL; AUTOMATED VEHICLES; CONTROL STRATEGY; MODEL; IMPACT;
D O I
10.1109/TITS.2024.3368474
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Deep reinforcement learning (DRL) algorithms often face challenges in achieving stability and efficiency due to significant policy gradient variance and inaccurate reward function estimation in complex scenarios. This study addresses these issues in the context of multi-objective car-following control tasks with time lag in traffic oscillations. We propose an expert demonstration reinforcement learning (EDRL) approach that aims to stabilize training, accelerate learning, and enhance car-following performance. The key idea is to leverage expert demonstrations, which represent superior car-following control experiences, to improve the DRL policy. Our method involves two sequential steps. In the first step, expert demonstrations are obtained during offline pretraining by utilizing prior traffic knowledge, including car-following trajectories from an empirical database and classic car-following models. In the second step, expert demonstrations are obtained during online training, where the agent interacts with the car-following environment. The EDRL agents are trained through supervised regression on the expert demonstrations using the behavioral cloning technique. Experimental results conducted in various traffic oscillation scenarios demonstrate that our proposed method significantly enhances training stability, learning speed, and rewards compared to baseline algorithms.
引用
收藏
页码:7751 / 7766
页数:16
相关论文
共 57 条
[11]   Efficient Deep Reinforcement Learning With Imitative Expert Priors for Autonomous Driving [J].
Huang, Zhiyu ;
Wu, Jingda ;
Lv, Chen .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (10) :7391-7403
[12]   Reinforcement Learning based cooperative longitudinal control for reducing traffic oscillations and improving platoon stability [J].
Jiang, Liming ;
Xie, Yuanchang ;
Evans, Nicholas G. ;
Wen, Xiao ;
Li, Tienan ;
Chen, Danjue .
TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2022, 141
[13]   Adaptive cruise control design for active congestion avoidance [J].
Kesting, Arne ;
Treiber, Martin ;
Schoenhof, Martin ;
Helbing, Dirk .
TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2008, 16 (06) :668-683
[14]   Enhanced intelligent driver model to access the impact of driving strategies on traffic capacity [J].
Kesting, Arne ;
Treiber, Martin ;
Helbing, Dirk .
PHILOSOPHICAL TRANSACTIONS OF THE ROYAL SOCIETY A-MATHEMATICAL PHYSICAL AND ENGINEERING SCIENCES, 2010, 368 (1928) :4585-4605
[15]   Modified DDPG car-following model with a real-world human driving experience with CARLA simulator [J].
Li, Dianzhao ;
Okhrin, Ostap .
TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2023, 147
[16]   Decision making of autonomous vehicles in lane change scenarios: Deep reinforcement learning approaches with risk awareness [J].
Li, Guofa ;
Yang, Yifan ;
Li, Shen ;
Qu, Xingda ;
Lyu, Nengchao ;
Li, Shengbo Eben .
TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2022, 134
[17]   Ecological Adaptive Cruise Control for Vehicles With Step-Gear Transmission Based on Reinforcement Learning [J].
Li, Guoqiang ;
Gorges, Daniel .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2020, 21 (11) :4895-4905
[18]   Anti-Disturbance Self-Supervised Reinforcement Learning for Perturbed Car-Following System [J].
Li, Meng ;
Li, Zhibin ;
Wang, Shunchao ;
Wang, Bingtong .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (09) :11318-11331
[19]   Enhancing Cooperation of Vehicle Merging Control in Heavy Traffic Using Communication-Based Soft Actor-Critic Algorithm [J].
Li, Meng ;
Li, Zhibin ;
Wang, Shunchao ;
Zheng, Si .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (06) :6491-6506
[20]   A Reinforcement Learning-Based Vehicle Platoon Control Strategy for Reducing Energy Consumption in Traffic Oscillations [J].
Li, Meng ;
Cao, Zehong ;
Li, Zhibin .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (12) :5309-5322