Safe model-based reinforcement learning for nonlinear optimal control with state and input constraints

被引:17
|
作者
Kim, Yeonsoo [1 ]
Kim, Jong Woo [2 ]
机构
[1] Kwangwoon Univ, Dept Chem Engn, 20 Kwangwoon Ro, Seoul 01897, South Korea
[2] Tech Univ Berlin, Chair Bioproc Engn, Berlin, Germany
基金
新加坡国家研究基金会;
关键词
approximate dynamic programming; barrier function; control Lyapunov function; reinforcement learning; Sontag's formula; PROGRAMS; SYSTEMS;
D O I
10.1002/aic.17601
中图分类号
TQ [化学工业];
学科分类号
0817 ;
摘要
Safety is a critical factor in reinforcement learning (RL) in chemical processes. In our previous work, we had proposed a new stability-guaranteed RL for unconstrained nonlinear control-affine systems. In the approximate policy iteration algorithm, a Lyapunov neural network (LNN) was updated while being restricted to the control Lyapunov function, and a policy was updated using a variation of Sontag's formula. In this study, we additionally consider state and input constraints by introducing a barrier function, and we extend the applicable type to general nonlinear systems. We augment the constraints into the objective function and use the LNN added with a Lyapunov barrier function to approximate the augmented value function. Sontag's formula input with this approximate function brings the states into its lower level set, thereby guaranteeing the constraints satisfaction and stability. We prove the practical asymptotic stability and forward invariance. The effectiveness is validated using four tank system simulations.
引用
收藏
页数:18
相关论文
共 50 条
  • [31] Model-Based Reinforcement Learning For Robot Control
    Li, Xiang
    Shang, Weiwei
    Cong, Shuang
    2020 5TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS (ICARM 2020), 2020, : 300 - 305
  • [32] Model-based reinforcement learning for approximate optimal regulation
    Kamalapurkar, Rushikesh
    Walters, Patrick
    Dixon, Warren E.
    AUTOMATICA, 2016, 64 : 94 - 104
  • [33] Model-Based Reinforcement Learning with a Generative Model is Minimax Optimal
    Agarwal, Alekh
    Kakade, Sham
    Yang, Lin F.
    CONFERENCE ON LEARNING THEORY, VOL 125, 2020, 125
  • [34] Control Lyapunov-barrier function-based safe reinforcement learning for nonlinear optimal control
    Wang, Yujia
    Wu, Zhe
    AICHE JOURNAL, 2024, 70 (03)
  • [35] Conservative and Adaptive Penalty for Model-Based Safe Reinforcement Learning
    Ma, Yecheng Jason
    Shen, Andrew
    Bastani, Osbert
    Jayaraman, Dinesh
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 5404 - 5412
  • [36] Safe Model-Based Reinforcement Learning for Systems With Parametric Uncertainties
    Mahmud, S. M. Nahid
    Nivison, Scott A.
    Bell, Zachary I.
    Kamalapurkar, Rushikesh
    FRONTIERS IN ROBOTICS AND AI, 2021, 8
  • [37] Reinforcement learning based robust tracking control for unmanned helicopter with state constraints and input saturation
    Feng, Yiting
    Zhou, Ye
    Ho, Hann Woei
    AEROSPACE SCIENCE AND TECHNOLOGY, 2024, 155
  • [38] Optimal Control for Constrained Discrete-Time Nonlinear Systems Based on Safe Reinforcement Learning
    Zhang, Lingzhi
    Xie, Lei
    Jiang, Yi
    Li, Zhishan
    Liu, Xueqin
    Su, Hongye
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (01) : 854 - 865
  • [39] Optimal Control for Constrained Discrete-Time Nonlinear Systems Based on Safe Reinforcement Learning
    Zhang, Lingzhi
    Xie, Lei
    Jiang, Yi
    Li, Zhishan
    Liu, Xueqin
    Su, Hongye
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (01) : 854 - 865
  • [40] Model-free Nearly Optimal Control of Constrained-Input Nonlinear Systems Based on Synchronous Reinforcement Learning
    Zhao, Han
    Guo, Lei
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 2162 - 2167