Robust Learning-Based Predictive Control for Discrete-Time Nonlinear Systems With Unknown Dynamics and State Constraints

被引:33
|
作者
Zhang, Xinglong [1 ]
Liu, Jiahang [1 ]
Xu, Xin [1 ]
Yu, Shuyou [2 ,3 ]
Chen, Hong [4 ]
机构
[1] Natl Univ Def Technol, Coll Intelligence Sci & Technol, Changsha 410073, Peoples R China
[2] Jilin Univ, State Key Lab Automot Simulat & Control, Changchun 130025, Peoples R China
[3] Jilin Univ, Dept Control Sci & Engn, Changchun 130025, Peoples R China
[4] Tongji Univ, Coll Elect & Informat Engn, Shanghai 201804, Peoples R China
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2022年 / 52卷 / 12期
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Robustness; Predictive control; Aerospace electronics; Computational modeling; Predictive models; Optimal control; Heuristic algorithms; Model predictive control (MPC); nonlinear systems; reinforcement learning (RL); robustness; state constraints; TRACKING CONTROL; LINEAR-SYSTEMS; STABILITY;
D O I
10.1109/TSMC.2022.3146284
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Robust model predictive control (MPC) is a well-known control technique for model-based control with constraints and uncertainties. In classic robust tube-based MPC approaches, an open-loop control sequence is computed via periodically solving an online nominal MPC problem, which requires prior model information and frequent access to onboard computational resources. In this article, we propose an efficient robust MPC solution based on receding horizon reinforcement learning, called r-LPC, for unknown nonlinear systems with state constraints and disturbances. The proposed r-LPC utilizes a Koopman operator-based prediction model obtained offline from precollected input-output datasets. Unlike classic tube-based MPC, in each prediction time interval of r-LPC, we use an actor-critic structure to learn a near-optimal feedback control policy rather than a control sequence. The resulting closed-loop control policy can be learned offline and deployed online or learned online in an asynchronous way. In the latter case, online learning can be activated whenever necessary; for instance, the safety constraint is violated with the deployed policy. The closed-loop recursive feasibility, robustness, and asymptotic stability are proven under function approximation errors of the actor-critic networks. Simulation and experimental results on two nonlinear systems with unknown dynamics and disturbances have demonstrated that our approach has better or comparable performance when compared with tube-based MPC and linear quadratic regulator, and outperforms a recently developed actor-critic learning approach.
引用
收藏
页码:7314 / 7327
页数:14
相关论文
共 50 条
  • [41] Resilient and robust finite-time H∞ control for uncertain discrete-time jump nonlinear systems
    Zhang, Yingqi
    Shi, Yan
    Shi, Peng
    APPLIED MATHEMATICAL MODELLING, 2017, 49 : 612 - 629
  • [42] Robust Fuzzy Model Predictive Control of Discrete-Time Takagi-Sugeno Systems With Nonlinear Local Models
    Teng, Long
    Wang, Youyi
    Cai, Wenjian
    Li, Hua
    IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2018, 26 (05) : 2915 - 2925
  • [43] Robust model predictive control for discrete-time nonlinear systems with delay-range-dependent
    Zhao, Jie-Mei
    Hu, Zhong-Hui
    Zhang, Li-Jun
    Kongzhi yu Juece/Control and Decision, 2015, 30 (01): : 59 - 64
  • [44] Robust Control for Discrete-Time Networked Control Systems
    Wu, Dongxiao
    Wu, Jun
    Chen, Sheng
    2010 8TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION (WCICA), 2010, : 3532 - 3537
  • [45] Reinforcement Learning-Based Optimal Stabilization for Unknown Nonlinear Systems Subject to Inputs With Uncertain Constraints
    Zhao, Bo
    Liu, Derong
    Luo, Chaomin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (10) : 4330 - 4340
  • [46] Robust adaptive iterative learning control for discrete-time nonlinear systems with both parametric and nonparametric uncertainties
    Yu, Miao
    Huang, Deqing
    He, Wei
    INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2016, 30 (07) : 972 - 985
  • [47] Direct adaptive robust NN control for a class of discrete-time nonlinear strict-feedback SISO systems
    Wen, Guo-Xing
    Liu, Yan-Jun
    Chen, C. L. Philip
    NEURAL COMPUTING & APPLICATIONS, 2012, 21 (06) : 1423 - 1431
  • [48] Robust Stabilization of Discrete-Time Uncertain Nonlinear Systems
    A. V. Savkin
    I. R. Petersen
    Journal of Optimization Theory and Applications, 1998, 96 : 87 - 107
  • [49] Robust Preview Control for a Class of Uncertain Discrete-Time Lipschitz Nonlinear Systems
    Yu, Xiao
    Liao, Fucheng
    Deng, Jiamei
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2018, 2018
  • [50] An ADDHP-based Q-learning algorithm for optimal tracking control of linear discrete-time systems with unknown dynamics
    Mu, Chaoxu
    Zhao, Qian
    Sun, Changyin
    Gao, Zhongke
    APPLIED SOFT COMPUTING, 2019, 82