Enhancing Car-Following Performance in Traffic Oscillations Using Expert Demonstration Reinforcement Learning

被引:3
|
作者
Li, Meng [1 ,2 ]
Li, Zhibin [1 ]
Cao, Zehong [3 ]
机构
[1] Southeast Univ, Sch Transportat, Nanjing 210096, Peoples R China
[2] Nanyang Technol Univ, Sch Mech & Aerosp Engn, Singapore 639798, Singapore
[3] Univ South Australia, STEM, Adelaide, SA 5095, Australia
基金
中国国家自然科学基金;
关键词
Training; Oscillators; Trajectory; Task analysis; Cloning; Safety; Databases; Expert demonstration; reinforcement learning; car-following control; traffic oscillation; ADAPTIVE CRUISE CONTROL; AUTOMATED VEHICLES; CONTROL STRATEGY; MODEL; IMPACT;
D O I
10.1109/TITS.2024.3368474
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Deep reinforcement learning (DRL) algorithms often face challenges in achieving stability and efficiency due to significant policy gradient variance and inaccurate reward function estimation in complex scenarios. This study addresses these issues in the context of multi-objective car-following control tasks with time lag in traffic oscillations. We propose an expert demonstration reinforcement learning (EDRL) approach that aims to stabilize training, accelerate learning, and enhance car-following performance. The key idea is to leverage expert demonstrations, which represent superior car-following control experiences, to improve the DRL policy. Our method involves two sequential steps. In the first step, expert demonstrations are obtained during offline pretraining by utilizing prior traffic knowledge, including car-following trajectories from an empirical database and classic car-following models. In the second step, expert demonstrations are obtained during online training, where the agent interacts with the car-following environment. The EDRL agents are trained through supervised regression on the expert demonstrations using the behavioral cloning technique. Experimental results conducted in various traffic oscillation scenarios demonstrate that our proposed method significantly enhances training stability, learning speed, and rewards compared to baseline algorithms.
引用
收藏
页码:7751 / 7766
页数:16
相关论文
共 50 条
  • [1] Multiagent Reinforcement Learning for Ecological Car-Following Control in Mixed Traffic
    Wang, Qun
    Ju, Fei
    Wang, Huaiyu
    Qian, Yahui
    Zhu, Meixin
    Zhuang, Weichao
    Wang, Liangmo
    IEEE TRANSACTIONS ON TRANSPORTATION ELECTRIFICATION, 2024, 10 (04): : 8671 - 8684
  • [2] A behavioral car-following model that captures traffic oscillations
    Chen, Danjue
    Laval, Jorge
    Zheng, Zuduo
    Ahn, Soyoung
    TRANSPORTATION RESEARCH PART B-METHODOLOGICAL, 2012, 46 (06) : 744 - 761
  • [3] Proactive Car-Following Using Deep-Reinforcement Learning
    Yen, Yi-Tung
    Chou, Jyun-Jhe
    Shih, Chi-Sheng
    Chen, Chih-Wei
    Tsung, Pei-Kuei
    2020 IEEE 23RD INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2020,
  • [4] MEASURING DRIVING PERFORMANCE BY CAR-FOLLOWING IN TRAFFIC
    BROOKHUIS, K
    DEWAARD, D
    MULDER, B
    ERGONOMICS, 1994, 37 (03) : 427 - 434
  • [5] A comparison study on the growth pattern of traffic oscillations in car-following experiments
    Zheng, Shi-Teng
    Jiang, Rui
    Tian, Junfang
    Li, Xiaopeng
    Jia, Bin
    Gao, Ziyou
    Yu, Shaowei
    TRANSPORTMETRICA B-TRANSPORT DYNAMICS, 2023, 11 (01) : 706 - 724
  • [6] Stability and performance of car-following models in congested traffic
    Aycin, MF
    Benekohal, RF
    JOURNAL OF TRANSPORTATION ENGINEERING, 2001, 127 (01) : 2 - 12
  • [7] Velocity control in car-following behavior with autonomous vehicles using reinforcement learning
    Wang, Zhe
    Huang, Helai
    Tang, Jinjun
    Meng, Xianwei
    Hu, Lipeng
    ACCIDENT ANALYSIS AND PREVENTION, 2022, 174
  • [8] A Car-following Control Algorithm Based on Deep Reinforcement Learning
    Zhu B.
    Jiang Y.-D.
    Zhao J.
    Chen H.
    Deng W.-W.
    Zhongguo Gonglu Xuebao/China Journal of Highway and Transport, 2019, 32 (06): : 53 - 60
  • [9] Towards robust car-following based on deep reinforcement learning
    Hart, Fabian
    Okhrin, Ostap
    Treiber, Martin
    TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2024, 159
  • [10] Driver Car-Following Model Based on Deep Reinforcement Learning
    Guo J.
    Li W.
    Luo Y.
    Chen T.
    Li K.
    Guo, Jinghua (guojh@xmu.edu.cn), 1600, SAE-China (43): : 571 - 579