Enhancing Car-Following Performance in Traffic Oscillations Using Expert Demonstration Reinforcement Learning

被引:3
|
作者
Li, Meng [1 ,2 ]
Li, Zhibin [1 ]
Cao, Zehong [3 ]
机构
[1] Southeast Univ, Sch Transportat, Nanjing 210096, Peoples R China
[2] Nanyang Technol Univ, Sch Mech & Aerosp Engn, Singapore 639798, Singapore
[3] Univ South Australia, STEM, Adelaide, SA 5095, Australia
基金
中国国家自然科学基金;
关键词
Training; Oscillators; Trajectory; Task analysis; Cloning; Safety; Databases; Expert demonstration; reinforcement learning; car-following control; traffic oscillation; ADAPTIVE CRUISE CONTROL; AUTOMATED VEHICLES; CONTROL STRATEGY; MODEL; IMPACT;
D O I
10.1109/TITS.2024.3368474
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Deep reinforcement learning (DRL) algorithms often face challenges in achieving stability and efficiency due to significant policy gradient variance and inaccurate reward function estimation in complex scenarios. This study addresses these issues in the context of multi-objective car-following control tasks with time lag in traffic oscillations. We propose an expert demonstration reinforcement learning (EDRL) approach that aims to stabilize training, accelerate learning, and enhance car-following performance. The key idea is to leverage expert demonstrations, which represent superior car-following control experiences, to improve the DRL policy. Our method involves two sequential steps. In the first step, expert demonstrations are obtained during offline pretraining by utilizing prior traffic knowledge, including car-following trajectories from an empirical database and classic car-following models. In the second step, expert demonstrations are obtained during online training, where the agent interacts with the car-following environment. The EDRL agents are trained through supervised regression on the expert demonstrations using the behavioral cloning technique. Experimental results conducted in various traffic oscillation scenarios demonstrate that our proposed method significantly enhances training stability, learning speed, and rewards compared to baseline algorithms.
引用
收藏
页码:7751 / 7766
页数:16
相关论文
共 50 条
  • [11] Dynamic Car-following Model Calibration with Deep Reinforcement Learning
    Naing, Htet
    Cai, Wentong
    Wu, Tiantian
    Yu, Liang
    2022 IEEE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2022, : 959 - 966
  • [12] Learning the Car-following Behavior of Drivers Using Maximum Entropy Deep Inverse Reinforcement Learning
    Zhou, Yang
    Fu, Rui
    Wang, Chang
    JOURNAL OF ADVANCED TRANSPORTATION, 2020, 2020
  • [13] CAR-FOLLOWING MODEL OF CONNECTED CRUISE CONTROL VEHICLES TO MITIGATE TRAFFIC OSCILLATIONS
    Qin, Yan-Yan
    Wang, Hao
    Chen, Quan
    Ran, Bin
    PROMET-TRAFFIC & TRANSPORTATION, 2019, 31 (06): : 603 - 610
  • [14] Car-following model for motorway traffic
    Wang, J
    Liu, RH
    Montgomery, F
    TRAFFIC FLOW THEORY 2005, 2005, (1934): : 33 - 42
  • [15] Enhancing Driver Car-Following Performance with a Distance and Acceleration Display
    Saffarian, Mehdi
    de Winter, Joost C. F.
    Happee, Riender
    IEEE TRANSACTIONS ON HUMAN-MACHINE SYSTEMS, 2013, 43 (01) : 8 - 16
  • [16] Car-following model of vehicular traffic
    Weng, YL
    Wu, TJ
    2001 INTERNATIONAL CONFERENCES ON INFO-TECH AND INFO-NET PROCEEDINGS, CONFERENCE A-G: INFO-TECH & INFO-NET: A KEY TO BETTER LIFE, 2001, : D101 - D106
  • [17] A car-following model for traffic flows
    Wang, Yong
    Proceedings of the 24th Chinese Control Conference, Vols 1 and 2, 2005, : 1638 - 1643
  • [18] Car-following models of vehicular traffic
    Weng, Yan-Lin
    Wu, Tie-Jun
    2002, Zhejiang University (03):
  • [19] Car-following models of vehicular traffic
    翁彦琳
    吴铁军
    Journal of Zhejiang University Science, 2002, (04) : 33 - 38
  • [20] Personalized Car-Following Control Based on a Hybrid of Reinforcement Learning and Supervised Learning
    Song, Dongjian
    Zhu, Bing
    Zhao, Jian
    Han, Jiayi
    Chen, Zhicheng
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (06) : 6014 - 6029