Towards Safe Decision-Making for Autonomous Vehicles at Unsignalized Intersections

被引:0
作者
Yang, Kai [1 ]
Li, Shen [2 ]
Chen, Yongli [1 ]
Cao, Dongpu [3 ]
Tang, Xiaolin [1 ]
机构
[1] Chongqing Univ, Coll Mech & Vehicle Engn, Chongqing 400044, Peoples R China
[2] Tsinghua Univ, Sch Civil Engn, Beijing 100084, Peoples R China
[3] Tsinghua Univ, Sch Vehicle & Mobil, Beijing 100084, Peoples R China
基金
中国国家自然科学基金;
关键词
Autonomous vehicles; decision-making safety; epistemic uncertainty; reinforcement learning; model predictive control;
D O I
10.1109/TVT.2024.3488749
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Urban autonomous driving decision-making poses a significant challenge, particularly when navigating unsignalized intersections. This complexity mainly stems from the stochastic interactions between various traffic participants. While reinforcement learning (RL)-based decision-making has shown promise, there are valid concerns regarding safety and adaptability. In particular, current RL-based models lack safeguards to prevent issuing potentially unsafe commands in unfamiliar scenarios that are not covered during training. To mitigate this issue, this paper proposes a safe decision-making framework to improve driving safety at unsignalized intersections. First, the RL-based policy is constructed based on the soft actor-critic (SAC) that maps environmental observations into actions directly. Subsequently, the reliability of the SAC policy is measured at run-time via epistemic uncertainty quantification. Furthermore, the risky actions of the RL policy are filtered based on the estimated reliability with integrating a risk-adaptive model predictive control (RAMPC) backup policy. Finally, an unsignalized intersection with occlusion is built via Simulation of Urban Mobility (SUMO). More importantly, several cases are carried out to simulate scenario data distribution shifts, i.e., traffic flow density variation, observation with sensor noise, and observation range decrease, which are not included in the RL policy training process. The results suggest that the proposed method can reduce risk and enhance the safety of autonomous driving at unsignalized intersections.
引用
收藏
页码:3830 / 3842
页数:13
相关论文
共 44 条
  • [1] Self-Learned Autonomous Driving at Unsignalized Intersections: A Hierarchical Reinforced Learning Approach for Feasible Decision-Making
    Al-Sharman, Mohammad
    Dempster, Rowan
    Daoud, Mohamed A.
    Nasr, Mahmoud
    Rayside, Derek
    Melek, William
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (11) : 12345 - 12356
  • [2] Deep Reinforcement Learning With NMPC Assistance Nash Switching for Urban Autonomous Driving
    Alighanbari, Sina
    Azad, Nasser L.
    [J]. IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2023, 8 (03): : 2604 - 2615
  • [3] An G, 2021, ADV NEUR IN
  • [4] Autonomous navigation at unsignalized intersections: A coupled reinforcement learning and model predictive control approach
    Bautista-Montesano, Rolando
    Galluzzi, Renato
    Ruan, Kangrui
    Fu, Yongjie
    Di, Xuan
    [J]. TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2022, 139
  • [5] Bouton Maxime, 2019, arXiv
  • [6] Learning Interaction-Aware Guidance for Trajectory Optimization in Dense Traffic Scenarios
    Brito, Bruno
    Agarwal, Achin
    Alonso-Mora, Javier
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (10) : 18808 - 18821
  • [7] Model Predictive Contouring Control for Collision Avoidance in Unstructured Dynamic Environments
    Brito, Bruno
    Floor, Boaz
    Ferranti, Laura
    Alonso-Mora, Javier
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2019, 4 (04) : 4459 - 4466
  • [8] Interaction aware cooperative trajectory planning for lane change maneuvers in dense traffic
    Burger, Christoph
    Schneider, Thomas
    Lauer, Martin
    [J]. 2020 IEEE 23RD INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2020,
  • [9] Continuous improvement of self-driving cars using dynamic confidence-aware reinforcement learning
    Cao, Zhong
    Jiang, Kun
    Zhou, Weitao
    Xu, Shaobing
    Peng, Huei
    Yang, Diange
    [J]. NATURE MACHINE INTELLIGENCE, 2023, 5 (02) : 145 - 158
  • [10] Trustworthy safety improvement for autonomous driving using reinforcement learning
    Cao, Zhong
    Xu, Shaobing
    Jiao, Xinyu
    Peng, Huei
    Yang, Diange
    [J]. TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2022, 138