Safe model-based reinforcement learning for nonlinear optimal control with state and input constraints

被引:17
|
作者
Kim, Yeonsoo [1 ]
Kim, Jong Woo [2 ]
机构
[1] Kwangwoon Univ, Dept Chem Engn, 20 Kwangwoon Ro, Seoul 01897, South Korea
[2] Tech Univ Berlin, Chair Bioproc Engn, Berlin, Germany
基金
新加坡国家研究基金会;
关键词
approximate dynamic programming; barrier function; control Lyapunov function; reinforcement learning; Sontag's formula; PROGRAMS; SYSTEMS;
D O I
10.1002/aic.17601
中图分类号
TQ [化学工业];
学科分类号
0817 ;
摘要
Safety is a critical factor in reinforcement learning (RL) in chemical processes. In our previous work, we had proposed a new stability-guaranteed RL for unconstrained nonlinear control-affine systems. In the approximate policy iteration algorithm, a Lyapunov neural network (LNN) was updated while being restricted to the control Lyapunov function, and a policy was updated using a variation of Sontag's formula. In this study, we additionally consider state and input constraints by introducing a barrier function, and we extend the applicable type to general nonlinear systems. We augment the constraints into the objective function and use the LNN added with a Lyapunov barrier function to approximate the augmented value function. Sontag's formula input with this approximate function brings the states into its lower level set, thereby guaranteeing the constraints satisfaction and stability. We prove the practical asymptotic stability and forward invariance. The effectiveness is validated using four tank system simulations.
引用
收藏
页数:18
相关论文
共 50 条
  • [41] Variable gain gradient descent-based reinforcement learning for robust optimal tracking control of uncertain nonlinear system with input constraints
    Mishra, Amardeep
    Ghosh, Satadal
    NONLINEAR DYNAMICS, 2022, 107 (03) : 2195 - 2214
  • [42] Variable gain gradient descent-based reinforcement learning for robust optimal tracking control of uncertain nonlinear system with input constraints
    Amardeep Mishra
    Satadal Ghosh
    Nonlinear Dynamics, 2022, 107 : 2195 - 2214
  • [43] Integral reinforcement learning-based optimal tracking control for uncertain nonlinear systems under input constraint and specified performance constraints
    Chang, Ru
    Liu, Zhi-Meng
    Li, Xiao-Bin
    Sun, Chang-Yin
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2024, 34 (13) : 8802 - 8824
  • [44] Safe Q-Learning for Data-Driven Nonlinear Optimal Control with Asymmetric State Constraints
    Zhao, Mingming
    Wang, Ding
    Song, Shijie
    Qiao, Junfei
    IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2024, 11 (12) : 2408 - 2422
  • [45] Safe Q-Learning for Data-Driven Nonlinear Optimal Control With Asymmetric State Constraints
    Mingming Zhao
    Ding Wang
    Shijie Song
    Junfei Qiao
    IEEE/CAA Journal of Automatica Sinica, 2024, 11 (12) : 2408 - 2422
  • [46] Robust safe reinforcement learning control of unknown continuous-time nonlinear systems with state constraints and disturbances
    Zhang, Haoran
    Zhao, Chunhui
    Ding, Jinliang
    JOURNAL OF PROCESS CONTROL, 2023, 128
  • [47] Control Approach Combining Reinforcement Learning and Model-Based Control
    Okawa, Yoshihiro
    Sasaki, Tomotake
    Iwane, Hidenao
    2019 12TH ASIAN CONTROL CONFERENCE (ASCC), 2019, : 1419 - 1424
  • [48] Input-to-state stability of model-based spacecraft formation control systems with communication constraints
    Liu, Xi
    Kumar, K. D.
    ACTA ASTRONAUTICA, 2011, 68 (11-12) : 1847 - 1859
  • [49] A model-based deep reinforcement learning method applied to finite-horizon optimal control of nonlinear control-affine system
    Kim, Jong Woo
    Park, Byung Jun
    Yoo, Haeun
    Oh, Tae Hoon
    Lee, Jay H.
    Lee, Jong Min
    JOURNAL OF PROCESS CONTROL, 2020, 87 (87) : 166 - 178
  • [50] Reinforcement Learning-Based Approximate Optimal Control for Attitude Reorientation Under State Constraints
    Dong, Hongyang
    Zhao, Xiaowei
    Yang, Haoyang
    IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2021, 29 (04) : 1664 - 1673