SGLPER: A safe end-to-end autonomous driving decision framework combining deep reinforcement learning and expert demonstrations via prioritized experience replay and the Gipps model☆

被引:0
作者
Cui, Jianping [1 ]
Yuan, Liang [1 ,2 ]
Xiao, Wendong [1 ]
Ran, Teng [1 ]
He, Li [1 ]
Zhang, Jianbo [1 ]
机构
[1] Xinjiang Univ, Sch Mech Engn, Urumqi, Peoples R China
[2] Shanghai Jiao Tong Univ, ICCI, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Bird's-eye-view; Expert demonstrations; Deep reinforcement learning; Autonomous driving; Collision avoidance;
D O I
10.1016/j.displa.2025.103041
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Despite significant advancements in deep reinforcement learning (DRL), existing methods for autonomous driving often need to overcome the cold-start problem, requiring extensive training to converge and fail to fully address safety concerns in dynamic driving environments. To address these limitations, we propose an efficient DRL framework, SGLPER, which integrates Prioritized Experience Replay (PER), expert demonstrations, and safe speed calculation model to improve learning efficiency and decision-making safety. Specifically, PER mitigates the cold-start problem by prioritizing high-value experiences and accelerating training convergence. The Long Short-Term Memory (LSTM) method also captures spatiotemporal information from observed states, enabling the agent to make informed decisions based on past experiences in complex, dynamic traffic scenarios. The safety strategy incorporates the Gipps model, introducing relatively safe speed limits into the reinforcement learning (RL) process to enhance driving safety. Moreover, Kullback-Leibler (KL) divergence combines RL with expert demonstrations, enabling the agent to learn human-like driving behaviors effectively. Experimental results in two simulated driving scenarios validate the robustness and effectiveness of the proposed framework. Compared to traditional DRL methods, SGLPER demonstrates safer strategies, higher success rates, and faster convergence. This study presents a promising approach for developing safer, more efficient autonomous driving systems.
引用
收藏
页数:10
相关论文
共 7 条
  • [1] End-to-End Autonomous Driving Decision Based on Deep Reinforcement Learning
    Huang Z.-Q.
    Qu Z.-W.
    Zhang J.
    Zhang Y.-X.
    Tian R.
    Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2020, 48 (09): : 1711 - 1719
  • [2] End-to-End Autonomous Driving Decision Based on Deep Reinforcement Learning
    Huang, Zhiqing
    Zhang, Ji
    Tian, Rui
    Zhang, Yanxin
    CONFERENCE PROCEEDINGS OF 2019 5TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND ROBOTICS (ICCAR), 2019, : 658 - 662
  • [3] Interpretable End-to-End Urban Autonomous Driving With Latent Deep Reinforcement Learning
    Chen, Jianyu
    Li, Shengbo Eben
    Tomizuka, Masayoshi
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (06) : 5068 - 5078
  • [4] SEAE: Stable end-to-end autonomous driving using event-triggered attention and exploration-driven deep reinforcement learning
    Cui, Jianping
    Yuan, Liang
    Xiao, Wendong
    Ran, Teng
    He, Li
    Zhang, Jianbo
    DISPLAYS, 2025, 87
  • [5] An End-to-End Deep Reinforcement Learning Based Modular Task Allocation Framework for Autonomous Mobile Systems
    Ma, Song
    Ruan, Jingqing
    Du, Yali
    Bucknall, Richard
    Liu, Yuanchang
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2025, 22 : 1519 - 1533
  • [6] An End-to-End Deep Reinforcement Learning Based Modular Task Allocation Framework for Autonomous Mobile Systems
    Ma, Song
    Ruan, Jingqing
    Du, Yali
    Bucknall, Richard
    Liu, Yuanchang
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2025, 22 : 1519 - 1533
  • [7] Towards end-to-end formation control for robotic fish via deep reinforcement learning with non-expert imitation?
    Sun, Yihao
    Yan, Chao
    Xiang, Xiaojia
    Zhou, Han
    Tang, Dengqing
    Zhu, Yi
    OCEAN ENGINEERING, 2023, 271