Safe Reinforcement Learning-based Driving Policy Design for Autonomous Vehicles on Highways

被引:4
作者
Nguyen, Hung Duy [1 ,2 ]
Han, Kyoungseok [1 ]
机构
[1] Kyungpook Natl Univ, Sch Mech Engn, Daegu 41566, South Korea
[2] TU Wien, Automat & Control Inst ACIN, A-1040 Vienna, Austria
基金
新加坡国家研究基金会;
关键词
Autonomous vehicles; collision avoidance; decision-making; finite state machine; safe reinforcement learning; DECISION-MAKING; ASSISTANCE; MODEL;
D O I
10.1007/s12555-023-0255-4
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Safe decision-making strategy of autonomous vehicles (AVs) plays a critical role in avoiding accidents. This study develops a safe reinforcement learning (safe-RL)-based driving policy for AVs on highways. The hierarchical framework is considered for the proposed safe-RL, where an upper layer executes a safe exploration-exploitation by modifying the exploring process of the epsilon-greedy algorithm, and a lower layer utilizes a finite state machine (FSM) approach to establish the safe conditions for state transitions. The proposed safe-RL-based driving policy improves the vehicle's safe driving ability using a Q-table that stores the values corresponding to each action state. Moreover, owing to the trade-off between the epsilon-greedy values and safe distance threshold, the simulation results demonstrate the superior performance of the proposed approach compared to other alternative RL approaches, such as the epsilon-greedy Q-learning (GQL) and decaying epsilon-greedy Q-learning (DGQL), in an uncertain traffic environment. This study's contributions are twofold: it improves the autonomous vehicle's exploration-exploitation and safe driving ability while utilizing the advantages of FSM when surrounding cars are inside safe-driving zones, and it analyzes the impact of safe-RL parameters in exploring the environment safely.
引用
收藏
页码:4098 / 4110
页数:13
相关论文
共 39 条
  • [1] [Anonymous], AI based Algorithms of Path Planning, Navigation and Control for Mobile Ground Robots and UAVs
  • [2] Baheri A, 2020, IEEE INT VEH SYM, P1550, DOI 10.1109/IV47402.2020.9304744
  • [3] Barber P., 1998, IEE C IND AUT CONTR
  • [4] Three Decades of Driver Assistance Systems Review and Future Perspectives
    Bengler, Klaus
    Dietmayer, Klaus
    Faerber, Berthold
    Maurer, Markus
    Stiller, Christoph
    Winner, Hermann
    [J]. IEEE INTELLIGENT TRANSPORTATION SYSTEMS MAGAZINE, 2014, 6 (04) : 6 - 22
  • [5] TESTING SOFTWARE DESIGN MODELED BY FINITE-STATE MACHINES
    CHOW, TS
    [J]. IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 1978, 4 (03) : 178 - 187
  • [6] An approach to tune fuzzy controllers based on reinforcement learning for autonomous vehicle control
    Dai, X
    Li, CK
    Rad, AB
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2005, 6 (03) : 285 - 293
  • [7] Dalal G, 2018, Arxiv, DOI arXiv:1801.08757
  • [8] Friedman J., 2001, ELEMENTS STAT LEARNI, V1
  • [9] García J, 2015, J MACH LEARN RES, V16, P1437
  • [10] Gu SD, 2024, Arxiv, DOI arXiv:2205.10330