Safe model-based reinforcement learning for nonlinear optimal control with state and input constraints

被引:17
|
作者
Kim, Yeonsoo [1 ]
Kim, Jong Woo [2 ]
机构
[1] Kwangwoon Univ, Dept Chem Engn, 20 Kwangwoon Ro, Seoul 01897, South Korea
[2] Tech Univ Berlin, Chair Bioproc Engn, Berlin, Germany
基金
新加坡国家研究基金会;
关键词
approximate dynamic programming; barrier function; control Lyapunov function; reinforcement learning; Sontag's formula; PROGRAMS; SYSTEMS;
D O I
10.1002/aic.17601
中图分类号
TQ [化学工业];
学科分类号
0817 ;
摘要
Safety is a critical factor in reinforcement learning (RL) in chemical processes. In our previous work, we had proposed a new stability-guaranteed RL for unconstrained nonlinear control-affine systems. In the approximate policy iteration algorithm, a Lyapunov neural network (LNN) was updated while being restricted to the control Lyapunov function, and a policy was updated using a variation of Sontag's formula. In this study, we additionally consider state and input constraints by introducing a barrier function, and we extend the applicable type to general nonlinear systems. We augment the constraints into the objective function and use the LNN added with a Lyapunov barrier function to approximate the augmented value function. Sontag's formula input with this approximate function brings the states into its lower level set, thereby guaranteeing the constraints satisfaction and stability. We prove the practical asymptotic stability and forward invariance. The effectiveness is validated using four tank system simulations.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Model-based reinforcement learning for nonlinear optimal control with practical asymptotic stability guarantees
    Kim, Yeonsoo
    Lee, Jong Min
    AICHE JOURNAL, 2020, 66 (10)
  • [2] Optimal control for a class of nonlinear systems with input constraints based on reinforcement learning
    Luo A.
    Xiao W.-B.
    Zhou Q.
    Lu R.-Q.
    Kongzhi Lilun Yu Yingyong/Control Theory and Applications, 2022, 39 (01): : 154 - 164
  • [3] Model-based safe reinforcement learning for nonlinear systems under uncertainty with constraints tightening approach
    Kim, Yeonsoo
    Oh, Tae Hoon
    COMPUTERS & CHEMICAL ENGINEERING, 2024, 183
  • [4] Safe reinforcement learning for affine nonlinear systems with state constraints and input saturation using control barrier functions
    Liu, Shihan
    Liu, Lijun
    Yu, Zhen
    NEUROCOMPUTING, 2023, 518 : 562 - 576
  • [5] Safe Transfer-Reinforcement-Learning-Based Optimal Control of Nonlinear Systems
    Wang, Yujia
    Xiao, Ming
    Wu, Zhe
    IEEE TRANSACTIONS ON CYBERNETICS, 2024, 54 (12) : 7272 - 7284
  • [6] Reinforcement learning-based output feedback control of nonlinear systems with input constraints
    He, P
    Jagannathan, S
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2005, 35 (01): : 150 - 154
  • [7] Model-Based Reinforcement Learning for Time-Optimal Velocity Control
    Hartmann, Gabriel
    Shiller, Zvi
    Azaria, Amos
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04): : 6185 - 6192
  • [8] A model-based deep reinforcement learning method applied to finite-horizon optimal control of nonlinear control-affine system
    Kim, Jong Woo
    Park, Byung Jun
    Yoo, Haeun
    Oh, Tae Hoon
    Lee, Jay H.
    Lee, Jong Min
    JOURNAL OF PROCESS CONTROL, 2020, 87 (87) : 166 - 178
  • [9] Fully cooperative games with state and input constraints using reinforcement learning based on control barrier functions
    Liu, Shihan
    Liu, Lijun
    Yu, Zhen
    ASIAN JOURNAL OF CONTROL, 2024, 26 (02) : 888 - 905
  • [10] Safe exploration in model-based reinforcement learning using control barrier functions
    Cohen, Max H.
    Belta, Calin
    AUTOMATICA, 2023, 147