Safe model-based reinforcement learning for nonlinear optimal control with state and input constraints

被引:17
|
作者
Kim, Yeonsoo [1 ]
Kim, Jong Woo [2 ]
机构
[1] Kwangwoon Univ, Dept Chem Engn, 20 Kwangwoon Ro, Seoul 01897, South Korea
[2] Tech Univ Berlin, Chair Bioproc Engn, Berlin, Germany
基金
新加坡国家研究基金会;
关键词
approximate dynamic programming; barrier function; control Lyapunov function; reinforcement learning; Sontag's formula; PROGRAMS; SYSTEMS;
D O I
10.1002/aic.17601
中图分类号
TQ [化学工业];
学科分类号
0817 ;
摘要
Safety is a critical factor in reinforcement learning (RL) in chemical processes. In our previous work, we had proposed a new stability-guaranteed RL for unconstrained nonlinear control-affine systems. In the approximate policy iteration algorithm, a Lyapunov neural network (LNN) was updated while being restricted to the control Lyapunov function, and a policy was updated using a variation of Sontag's formula. In this study, we additionally consider state and input constraints by introducing a barrier function, and we extend the applicable type to general nonlinear systems. We augment the constraints into the objective function and use the LNN added with a Lyapunov barrier function to approximate the augmented value function. Sontag's formula input with this approximate function brings the states into its lower level set, thereby guaranteeing the constraints satisfaction and stability. We prove the practical asymptotic stability and forward invariance. The effectiveness is validated using four tank system simulations.
引用
收藏
页数:18
相关论文
共 50 条
  • [11] Model-Based Safe Reinforcement Learning With Time-Varying Constraints: Applications to Intelligent Vehicles
    Zhang, Xinglong
    Peng, Yaoqian
    Luo, Biao
    Pan, Wei
    Xu, Xin
    Xie, Haibin
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2024, 71 (10) : 12744 - 12753
  • [12] Control Lyapunov-barrier function-based safe reinforcement learning for nonlinear optimal control
    Wang, Yujia
    Wu, Zhe
    AICHE JOURNAL, 2024, 70 (03)
  • [13] Reinforcement learning based robust tracking control for unmanned helicopter with state constraints and input saturation
    Feng, Yiting
    Zhou, Ye
    Ho, Hann Woei
    AEROSPACE SCIENCE AND TECHNOLOGY, 2024, 155
  • [14] Advances in model-based reinforcement learning for Adaptive Optics control
    Nousiainen, Jalo
    Engler, Byron
    Kasper, Markus
    Helin, Tapio
    Heritier, Cedric T.
    Rajani, Chang
    ADAPTIVE OPTICS SYSTEMS VIII, 2022, 12185
  • [15] Integral reinforcement learning-based optimal tracking control for uncertain nonlinear systems under input constraint and specified performance constraints
    Chang, Ru
    Liu, Zhi-Meng
    Li, Xiao-Bin
    Sun, Chang-Yin
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2024, 34 (13) : 8802 - 8824
  • [16] Temporal logic guided safe model-based reinforcement learning: A hybrid systems approach
    Cohen, Max H.
    Serlin, Zachary
    Leahy, Kevin
    Belta, Calin
    NONLINEAR ANALYSIS-HYBRID SYSTEMS, 2023, 47
  • [17] Off-policy model-based end-to-end safe reinforcement learning
    Kanso, Soha
    Jha, Mayank Shekhar
    Theilliol, Didier
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2024, 34 (04) : 2806 - 2831
  • [18] Reinforcement Learning-Based Approximate Optimal Control for Attitude Reorientation Under State Constraints
    Dong, Hongyang
    Zhao, Xiaowei
    Yang, Haoyang
    IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2021, 29 (04) : 1664 - 1673
  • [19] Cooperative Model-Based Reinforcement Learning for Approximate Optimal Tracking
    Greene, Max L.
    Bell, Zachary, I
    Nivison, Scott A.
    How, Jonathan P.
    Dixon, Warren E.
    2021 AMERICAN CONTROL CONFERENCE (ACC), 2021, : 1973 - 1978
  • [20] Laboratory experiments of model-based reinforcement learning for adaptive optics control
    Nousiainen, Jalo
    Engler, Byron
    Kasper, Markus
    Rajani, Chang
    Helin, Tapio
    Heritier, Cedric T.
    Quanz, Sascha P.
    Glauser, Adrian M.
    JOURNAL OF ASTRONOMICAL TELESCOPES INSTRUMENTS AND SYSTEMS, 2024, 10 (01)