Safe Reinforcement Learning for Model-Reference Trajectory Tracking of Uncertain Autonomous Vehicles With Model-Based Acceleration

被引:23
作者
Hu, Yifan [1 ]
Fu, Junjie [1 ,2 ]
Wen, Guanghui [1 ]
机构
[1] Southeast Univ, Sch Math, Nanjing 210096, Peoples R China
[2] Purple Mt Labs, Nanjing 211111, Peoples R China
来源
IEEE TRANSACTIONS ON INTELLIGENT VEHICLES | 2023年 / 8卷 / 03期
基金
中国国家自然科学基金;
关键词
Safety; Predictive models; Trajectory tracking; Training; Reinforcement learning; Heuristic algorithms; Uncertainty; Model-reference control; autonomous vehicle; safe reinforcement learning; model-based reinforcement learning; Gaussian process; control barrier function;
D O I
10.1109/TIV.2022.3233592
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Applying reinforcement learning (RL) algorithms to control systems design remains a challenging task due to the potential unsafe exploration and the low sample efficiency. In this paper, we propose a novel safe model-based RL algorithm to solve the collision-free model-reference trajectory tracking problem of uncertain autonomous vehicles (AVs). Firstly, a new type of robust control barrier function (CBF) condition for collision-avoidance is derived for the uncertain AVs by incorporating the estimation of the system uncertainty with Gaussian process (GP) regression. Then, a robust CBF-based RL control structure is proposed, where the nominal control input is composed of the RL policy and a model-based reference control policy. The actual control input obtained from the quadratic programming problem can satisfy the constraints of collision-avoidance, input saturation and velocity boundedness simultaneously with a relatively high probability. Finally, within this control structure, a Dyna-style safe model-based RL algorithm is proposed, where the safe exploration is achieved through executing the robust CBF-based actions and the sample efficiency is improved by leveraging the GP models. The superior learning performance of the proposed RL control structure is demonstrated through simulation experiments.
引用
收藏
页码:2332 / 2344
页数:13
相关论文
共 50 条
  • [41] Robust MPC-based trajectory tracking of autonomous underwater vehicles with model uncertainty
    Yan, Zheping
    Yan, Jinyu
    Cai, Sijia
    Yu, Yuyang
    Wu, Yifan
    OCEAN ENGINEERING, 2023, 286
  • [42] Model-based safe reinforcement learning for nonlinear systems under uncertainty with constraints tightening approach
    Kim, Yeonsoo
    Oh, Tae Hoon
    COMPUTERS & CHEMICAL ENGINEERING, 2024, 183
  • [43] Model-Based Reinforcement Learning for Physical Systems Without Velocity and Acceleration Measurements
    Dalla Libera, Alberto
    Romeres, Diego
    Jha, Devesh K.
    Yerazunis, Bill
    Nikovski, Daniel
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (02) : 3548 - 3555
  • [44] Uncertainty-Aware Contact-Safe Model-Based Reinforcement Learning
    Kuo, Cheng-Yu
    Schaarschmidt, Andreas
    Cui, Yunduan
    Asfour, Tamim
    Matsubara, Takamitsu
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) : 3918 - 3925
  • [45] Incremental model-based reinforcement learning with model constraint
    Yang, Zhiyou
    Fu, Mingsheng
    Qu, Hong
    Li, Fan
    Shi, Shuqing
    Hu, Wang
    NEURAL NETWORKS, 2025, 185
  • [46] Model-Based Reinforcement Learning for Cavity Filter Tuning
    Nimara, Doumitrou Daniil
    Malek-Mohammadi, Mohammadreza
    Wei, Jieqiang
    Huang, Vincent
    Ogren, Petter
    LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 211, 2023, 211
  • [47] Learnable Weighting Mechanism in Model-based Reinforcement Learning
    Huang W.-Z.
    Yin Q.-Y.
    Zhang J.-G.
    Huang K.-Q.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (06): : 2765 - 2775
  • [48] Safe exploration in model-based reinforcement learning using control barrier functions
    Cohen, Max H.
    Belta, Calin
    AUTOMATICA, 2023, 147
  • [49] Offline Model-Based Reinforcement Learning for Tokamak Control
    Char, Ian
    Abbate, Joseph
    Bardoczi, Laszlo
    Boyer, Mark D.
    Chung, Youngseog
    Conlin, Rory
    Erickson, Keith
    Mehta, Viraj
    Richner, Nathan
    Kolemen, Egemen
    Schneider, Jeff
    LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 211, 2023, 211
  • [50] Barrier Lyapunov Function-Based Safe Reinforcement Learning for Autonomous Vehicles With Optimized Backstepping
    Zhang, Yuxiang
    Liang, Xiaoling
    Li, Dongyu
    Ge, Shuzhi Sam
    Gao, Bingzhao
    Chen, Hong
    Lee, Tong Heng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (02) : 2066 - 2080