Safe model-based reinforcement learning for nonlinear optimal control with state and input constraints

被引:17
|
作者
Kim, Yeonsoo [1 ]
Kim, Jong Woo [2 ]
机构
[1] Kwangwoon Univ, Dept Chem Engn, 20 Kwangwoon Ro, Seoul 01897, South Korea
[2] Tech Univ Berlin, Chair Bioproc Engn, Berlin, Germany
基金
新加坡国家研究基金会;
关键词
approximate dynamic programming; barrier function; control Lyapunov function; reinforcement learning; Sontag's formula; PROGRAMS; SYSTEMS;
D O I
10.1002/aic.17601
中图分类号
TQ [化学工业];
学科分类号
0817 ;
摘要
Safety is a critical factor in reinforcement learning (RL) in chemical processes. In our previous work, we had proposed a new stability-guaranteed RL for unconstrained nonlinear control-affine systems. In the approximate policy iteration algorithm, a Lyapunov neural network (LNN) was updated while being restricted to the control Lyapunov function, and a policy was updated using a variation of Sontag's formula. In this study, we additionally consider state and input constraints by introducing a barrier function, and we extend the applicable type to general nonlinear systems. We augment the constraints into the objective function and use the LNN added with a Lyapunov barrier function to approximate the augmented value function. Sontag's formula input with this approximate function brings the states into its lower level set, thereby guaranteeing the constraints satisfaction and stability. We prove the practical asymptotic stability and forward invariance. The effectiveness is validated using four tank system simulations.
引用
收藏
页数:18
相关论文
共 50 条
  • [11] Machine learning model-based optimal tracking control of nonlinear affine systems with safety constraints
    Wang, Yujia
    Wu, Zhe
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2025, 35 (02) : 511 - 535
  • [12] Safe Transfer-Reinforcement-Learning-Based Optimal Control of Nonlinear Systems
    Wang, Yujia
    Xiao, Ming
    Wu, Zhe
    IEEE TRANSACTIONS ON CYBERNETICS, 2024, 54 (12) : 7272 - 7284
  • [13] Reinforcement learning-based output feedback control of nonlinear systems with input constraints
    He, P
    Jagannathan, S
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2005, 35 (01): : 150 - 154
  • [14] Reinforcement learning-based output feedback control of nonlinear systems with input constraints
    He, P
    Jagannathan, S
    PROCEEDINGS OF THE 2004 AMERICAN CONTROL CONFERENCE, VOLS 1-6, 2004, : 2563 - 2568
  • [15] Model-Based Reinforcement Learning for Optimal Feedback Control of Switched Systems
    Greene, Max L.
    Abudia, Moad
    Kamalapurkar, Rushikesh
    Dixon, Warren E.
    2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 162 - 167
  • [16] Efficient model-based reinforcement learning for approximate online optimal control
    Kamalapurkar, Rushikesh
    Rosenfeld, Joel A.
    Dixon, Warren E.
    AUTOMATICA, 2016, 74 : 247 - 258
  • [17] Model-Based Cross-Scale Reinforcement Learning Optimal Control
    Li, Gonghe
    Zhou, Linna
    Liu, Xiaomin
    Yang, Chunyu
    2024 6th International Conference on Electronic Engineering and Informatics, EEI 2024, 2024, : 906 - 910
  • [18] Model-Based Reinforcement Learning for Time-Optimal Velocity Control
    Hartmann, Gabriel
    Shiller, Zvi
    Azaria, Amos
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04): : 6185 - 6192
  • [19] Safe exploration in model-based reinforcement learning using control barrier functions
    Cohen, Max H.
    Belta, Calin
    AUTOMATICA, 2023, 147
  • [20] Safe Model-based Reinforcement Learning with Stability Guarantees
    Berkenkamp, Felix
    Turchetta, Matteo
    Schoellig, Angela P.
    Krause, Andreas
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30