Enhancing Car-Following Performance in Traffic Oscillations Using Expert Demonstration Reinforcement Learning

被引:3
|
作者
Li, Meng [1 ,2 ]
Li, Zhibin [1 ]
Cao, Zehong [3 ]
机构
[1] Southeast Univ, Sch Transportat, Nanjing 210096, Peoples R China
[2] Nanyang Technol Univ, Sch Mech & Aerosp Engn, Singapore 639798, Singapore
[3] Univ South Australia, STEM, Adelaide, SA 5095, Australia
基金
中国国家自然科学基金;
关键词
Training; Oscillators; Trajectory; Task analysis; Cloning; Safety; Databases; Expert demonstration; reinforcement learning; car-following control; traffic oscillation; ADAPTIVE CRUISE CONTROL; AUTOMATED VEHICLES; CONTROL STRATEGY; MODEL; IMPACT;
D O I
10.1109/TITS.2024.3368474
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Deep reinforcement learning (DRL) algorithms often face challenges in achieving stability and efficiency due to significant policy gradient variance and inaccurate reward function estimation in complex scenarios. This study addresses these issues in the context of multi-objective car-following control tasks with time lag in traffic oscillations. We propose an expert demonstration reinforcement learning (EDRL) approach that aims to stabilize training, accelerate learning, and enhance car-following performance. The key idea is to leverage expert demonstrations, which represent superior car-following control experiences, to improve the DRL policy. Our method involves two sequential steps. In the first step, expert demonstrations are obtained during offline pretraining by utilizing prior traffic knowledge, including car-following trajectories from an empirical database and classic car-following models. In the second step, expert demonstrations are obtained during online training, where the agent interacts with the car-following environment. The EDRL agents are trained through supervised regression on the expert demonstrations using the behavioral cloning technique. Experimental results conducted in various traffic oscillation scenarios demonstrate that our proposed method significantly enhances training stability, learning speed, and rewards compared to baseline algorithms.
引用
收藏
页码:7751 / 7766
页数:16
相关论文
共 50 条
  • [31] Reinforcement Learning-based Car-Following Control for Autonomous Vehicles with OTFS
    Liu, Yulin
    Shi, Yuye
    Zhang, Xiaoqi
    Wu, Jun
    Yang, Songyuan
    2024 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC 2024, 2024,
  • [32] A Reinforcement Learning Framework for Video Frame-Based Autonomous Car-Following
    Masmoudi, Mehdi
    Friji, Hamdi
    Ghazzai, Hakim
    Massoud, Yehia
    IEEE OPEN JOURNAL OF INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 2 : 111 - 127
  • [33] Human-like autonomous car-following model with deep reinforcement learning
    Zhu, Meixin
    Wang, Xuesong
    Wang, Yinhai
    TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2018, 97 : 348 - 368
  • [34] Deep Reinforcement Learning Car-Following Model Considering Longitudinal and Lateral Control
    Qin, Pinpin
    Tan, Hongyun
    Li, Hao
    Wen, Xuguang
    SUSTAINABILITY, 2022, 14 (24)
  • [35] Simulating train movement in railway traffic using a car-following model
    李克平
    管立加
    Chinese Physics B, 2009, 18 (06) : 2200 - 2204
  • [36] Simulating train movement in railway traffic using a car-following model
    Li Ke-Ping
    Guan Li-Jia
    CHINESE PHYSICS B, 2009, 18 (06) : 2200 - 2204
  • [37] Modeling the Effects of Autonomous Vehicles on Human Driver Car-Following Behaviors Using Inverse Reinforcement Learning
    Wen, Xiao
    Jian, Sisi
    He, Dengbo
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (12) : 13903 - 13915
  • [38] Car-following strategy of intelligent connected vehicle using extended disturbance observer adjusted by reinforcement learning
    Yan, Ruidong
    Li, Penghui
    Gao, Hongbo
    Huang, Jin
    Wang, Chengbo
    CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2024, 9 (02) : 365 - 373
  • [39] A parsimonious model for the formation of oscillations in car-following models
    Laval, Jorge A.
    Toth, Christopher S.
    Zhou, Yi
    TRANSPORTATION RESEARCH PART B-METHODOLOGICAL, 2014, 70 : 228 - 238
  • [40] Car-following Behavior Model Learning Using Timed Automata
    Zhang, Yihuan
    Lin, Qin
    Wang, Jun
    Verwer, Sicco
    IFAC PAPERSONLINE, 2017, 50 (01): : 2353 - 2358