Enhancing Car-Following Performance in Traffic Oscillations Using Expert Demonstration Reinforcement Learning

被引:3
|
作者
Li, Meng [1 ,2 ]
Li, Zhibin [1 ]
Cao, Zehong [3 ]
机构
[1] Southeast Univ, Sch Transportat, Nanjing 210096, Peoples R China
[2] Nanyang Technol Univ, Sch Mech & Aerosp Engn, Singapore 639798, Singapore
[3] Univ South Australia, STEM, Adelaide, SA 5095, Australia
基金
中国国家自然科学基金;
关键词
Training; Oscillators; Trajectory; Task analysis; Cloning; Safety; Databases; Expert demonstration; reinforcement learning; car-following control; traffic oscillation; ADAPTIVE CRUISE CONTROL; AUTOMATED VEHICLES; CONTROL STRATEGY; MODEL; IMPACT;
D O I
10.1109/TITS.2024.3368474
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Deep reinforcement learning (DRL) algorithms often face challenges in achieving stability and efficiency due to significant policy gradient variance and inaccurate reward function estimation in complex scenarios. This study addresses these issues in the context of multi-objective car-following control tasks with time lag in traffic oscillations. We propose an expert demonstration reinforcement learning (EDRL) approach that aims to stabilize training, accelerate learning, and enhance car-following performance. The key idea is to leverage expert demonstrations, which represent superior car-following control experiences, to improve the DRL policy. Our method involves two sequential steps. In the first step, expert demonstrations are obtained during offline pretraining by utilizing prior traffic knowledge, including car-following trajectories from an empirical database and classic car-following models. In the second step, expert demonstrations are obtained during online training, where the agent interacts with the car-following environment. The EDRL agents are trained through supervised regression on the expert demonstrations using the behavioral cloning technique. Experimental results conducted in various traffic oscillation scenarios demonstrate that our proposed method significantly enhances training stability, learning speed, and rewards compared to baseline algorithms.
引用
收藏
页码:7751 / 7766
页数:16
相关论文
共 50 条
  • [21] Car-following models of vehicular traffic
    Weng Yan-lin
    Wu Tie-jun
    Journal of Zhejiang University-SCIENCE A, 2002, 3 (4): : 412 - 417
  • [22] Traffic Oscillations Mitigation in Vehicle Platoon Using a Car-Following Control Model for Connected and Autonomous Vehicle
    Gao, Zhijun
    Wang, Jiangfeng
    Zhang, Xi
    Dong, Jiakuan
    Chen, Lei
    Yan, Xuedong
    Zhang, Wei
    JOURNAL OF ADVANCED TRANSPORTATION, 2019, 2019
  • [23] Improved deep reinforcement learning for car-following decision-making
    Yang, Xiaoxue
    Zou, Yajie
    Zhang, Hao
    Qu, Xiaobo
    Chen, Lei
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2023, 624
  • [24] Longitudinal Dynamic versus Kinematic Models for Car-Following Control Using Deep Reinforcement Learning
    Lin, Yuan
    McPhee, John
    Azad, Nasser L.
    2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2019, : 1504 - 1510
  • [25] Traffic Experiment Reveals the Nature of Car-Following
    Jiang, Rui
    Hu, Mao-Bin
    Zhang, H. M.
    Gao, Zi-You
    Jia, Bin
    Wu, Qing-Song
    Wang, Bing
    Yang, Ming
    PLOS ONE, 2014, 9 (04):
  • [26] An improved car-following model for railway traffic
    Li, KePing
    Gao, ZiYou
    JOURNAL OF ADVANCED TRANSPORTATION, 2013, 47 (04) : 475 - 482
  • [27] A mixed traffic car-following behavior model
    Wang, Xinke
    Zhang, Jian
    Li, Honghai
    He, Zhengbing
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2023, 632
  • [28] Deep Reinforcement Learning Car-Following Control Based on Multivehicle Motion Prediction
    Wang, Tao
    Qu, Dayi
    Wang, Kedong
    Dai, Shouchen
    ELECTRONICS, 2024, 13 (06)
  • [29] Car-Following Behavior Modeling With Maximum Entropy Deep Inverse Reinforcement Learning
    Nan, Jiangfeng
    Deng, Weiwen
    Zhang, Ruzheng
    Zhao, Rui
    Wang, Ying
    Ding, Juan
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2024, 9 (02): : 3998 - 4010
  • [30] Improving Car-Following Control in Mixed Traffic: A Deep Reinforcement Learning Framework with Aggregated Human-Driven Vehicles
    Chen, Xianda
    Tiu, PakHin
    Zhang, Yihuai
    Zhu, Meixin
    Zheng, Xinhu
    Wang, Yinhai
    2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024, 2024, : 627 - 632