Safe model-based reinforcement learning for nonlinear optimal control with state and input constraints

被引:17
|
作者
Kim, Yeonsoo [1 ]
Kim, Jong Woo [2 ]
机构
[1] Kwangwoon Univ, Dept Chem Engn, 20 Kwangwoon Ro, Seoul 01897, South Korea
[2] Tech Univ Berlin, Chair Bioproc Engn, Berlin, Germany
基金
新加坡国家研究基金会;
关键词
approximate dynamic programming; barrier function; control Lyapunov function; reinforcement learning; Sontag's formula; PROGRAMS; SYSTEMS;
D O I
10.1002/aic.17601
中图分类号
TQ [化学工业];
学科分类号
0817 ;
摘要
Safety is a critical factor in reinforcement learning (RL) in chemical processes. In our previous work, we had proposed a new stability-guaranteed RL for unconstrained nonlinear control-affine systems. In the approximate policy iteration algorithm, a Lyapunov neural network (LNN) was updated while being restricted to the control Lyapunov function, and a policy was updated using a variation of Sontag's formula. In this study, we additionally consider state and input constraints by introducing a barrier function, and we extend the applicable type to general nonlinear systems. We augment the constraints into the objective function and use the LNN added with a Lyapunov barrier function to approximate the augmented value function. Sontag's formula input with this approximate function brings the states into its lower level set, thereby guaranteeing the constraints satisfaction and stability. We prove the practical asymptotic stability and forward invariance. The effectiveness is validated using four tank system simulations.
引用
收藏
页数:18
相关论文
共 50 条
  • [21] SAMBA: safe model-based & active reinforcement learning
    Alexander I. Cowen-Rivers
    Daniel Palenicek
    Vincent Moens
    Mohammed Amin Abdullah
    Aivar Sootla
    Jun Wang
    Haitham Bou-Ammar
    Machine Learning, 2022, 111 : 173 - 203
  • [22] SAMBA: safe model-based & active reinforcement learning
    Cowen-Rivers, Alexander, I
    Palenicek, Daniel
    Moens, Vincent
    Abdullah, Mohammed Amin
    Sootla, Aivar
    Wang, Jun
    Bou-Ammar, Haitham
    MACHINE LEARNING, 2022, 111 (01) : 173 - 203
  • [23] Safe Robot Execution in Model-Based Reinforcement Learning
    Martinez, David
    Alenya, Guillem
    Torras, Carme
    2015 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2015, : 6422 - 6427
  • [24] Reinforcement learning for adaptive optimal control of unknown continuous-time nonlinear systems with input constraints
    Yang, Xiong
    Liu, Derong
    Wang, Ding
    INTERNATIONAL JOURNAL OF CONTROL, 2014, 87 (03) : 553 - 566
  • [25] Model-Based Safe Reinforcement Learning With Time-Varying Constraints: Applications to Intelligent Vehicles
    Zhang, Xinglong
    Peng, Yaoqian
    Luo, Biao
    Pan, Wei
    Xu, Xin
    Xie, Haibin
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2024, 71 (10) : 12744 - 12753
  • [26] Model-Based Safe Reinforcement Learning With Time-Varying Constraints: Applications to Intelligent Vehicles
    Zhang, Xinglong
    Peng, Yaoqian
    Luo, Biao
    Pan, Wei
    Xu, Xin
    Xie, Haibin
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2024, 71 (10) : 12744 - 12753
  • [27] Reinforcement-Learning-Based Adaptive Optimal Flight Control with Output Feedback and Input Constraints
    Sun, Bo
    van Kampen, Erik-Jan
    JOURNAL OF GUIDANCE CONTROL AND DYNAMICS, 2021, 44 (09) : 1685 - 1691
  • [28] Model-Based Reinforcement Learning for Approximate Optimal Control with Temporal Logic Specifications
    Cohen, Max H.
    Belta, Cahn
    HSCC2021: PROCEEDINGS OF THE 24TH INTERNATIONAL CONFERENCE ON HYBRID SYSTEMS: COMPUTATION AND CONTROL (PART OF CPS-IOT WEEK), 2021,
  • [29] A Model-Based Reinforcement Learning Approach to Time-Optimal Control Problems
    Liao, Hsuan-Cheng
    Liu, Jing-Sin
    ADVANCES AND TRENDS IN ARTIFICIAL INTELLIGENCE: FROM THEORY TO PRACTICE, 2019, 11606 : 657 - 665
  • [30] A Safe and Data-Efficient Model-Based Reinforcement Learning System for HVAC Control
    Ding, Xianzhong
    An, Zhiyu
    Rathee, Arya
    Du, Wan
    IEEE INTERNET OF THINGS JOURNAL, 2025, 12 (07): : 8014 - 8032