Robust Learning-Based Predictive Control for Discrete-Time Nonlinear Systems With Unknown Dynamics and State Constraints

被引:33
|
作者
Zhang, Xinglong [1 ]
Liu, Jiahang [1 ]
Xu, Xin [1 ]
Yu, Shuyou [2 ,3 ]
Chen, Hong [4 ]
机构
[1] Natl Univ Def Technol, Coll Intelligence Sci & Technol, Changsha 410073, Peoples R China
[2] Jilin Univ, State Key Lab Automot Simulat & Control, Changchun 130025, Peoples R China
[3] Jilin Univ, Dept Control Sci & Engn, Changchun 130025, Peoples R China
[4] Tongji Univ, Coll Elect & Informat Engn, Shanghai 201804, Peoples R China
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2022年 / 52卷 / 12期
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Robustness; Predictive control; Aerospace electronics; Computational modeling; Predictive models; Optimal control; Heuristic algorithms; Model predictive control (MPC); nonlinear systems; reinforcement learning (RL); robustness; state constraints; TRACKING CONTROL; LINEAR-SYSTEMS; STABILITY;
D O I
10.1109/TSMC.2022.3146284
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Robust model predictive control (MPC) is a well-known control technique for model-based control with constraints and uncertainties. In classic robust tube-based MPC approaches, an open-loop control sequence is computed via periodically solving an online nominal MPC problem, which requires prior model information and frequent access to onboard computational resources. In this article, we propose an efficient robust MPC solution based on receding horizon reinforcement learning, called r-LPC, for unknown nonlinear systems with state constraints and disturbances. The proposed r-LPC utilizes a Koopman operator-based prediction model obtained offline from precollected input-output datasets. Unlike classic tube-based MPC, in each prediction time interval of r-LPC, we use an actor-critic structure to learn a near-optimal feedback control policy rather than a control sequence. The resulting closed-loop control policy can be learned offline and deployed online or learned online in an asynchronous way. In the latter case, online learning can be activated whenever necessary; for instance, the safety constraint is violated with the deployed policy. The closed-loop recursive feasibility, robustness, and asymptotic stability are proven under function approximation errors of the actor-critic networks. Simulation and experimental results on two nonlinear systems with unknown dynamics and disturbances have demonstrated that our approach has better or comparable performance when compared with tube-based MPC and linear quadratic regulator, and outperforms a recently developed actor-critic learning approach.
引用
收藏
页码:7314 / 7327
页数:14
相关论文
共 50 条
  • [1] Learning-Based Predictive Control for Discrete-Time Nonlinear Systems With Stochastic Disturbances
    Xu, Xin
    Chen, Hong
    Lian, Chuanqiang
    Li, Dazi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (12) : 6202 - 6213
  • [2] Learning-based robust output tracking control for unknown discrete-time nonlinear systems with dynamic uncertainty
    Liu, Fang
    Peng, Hui
    NEUROCOMPUTING, 2024, 606
  • [3] Reinforcement Learning-Based Model Predictive Control for Discrete-Time Systems
    Lin, Min
    Sun, Zhongqi
    Xia, Yuanqing
    Zhang, Jinhui
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 3312 - 3324
  • [4] H∞ control for a class of discrete-time switched systems with state constraints
    Su, Qingyu
    Zhao, Jun
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART I-JOURNAL OF SYSTEMS AND CONTROL ENGINEERING, 2012, 226 (I10) : 1301 - 1310
  • [5] Robust Adaptive Control for Stochastic Discrete-Time Nonlinear Systems and Application to Gas Engine as an Electric Vehicle Extender
    Yang, Jun
    Zhang, Qinglin
    Li, Yanxiao
    Wang, Jian
    IEEE ACCESS, 2020, 8 : 156433 - 156441
  • [6] Robust fuzzy model predictive control for nonlinear discrete-time systems
    Mahmoudabadi, Parvin
    Naderi Akhormeh, Alireza
    INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2024, 38 (03) : 938 - 953
  • [7] Robust Learning-Based Model Predictive Control for Intelligent Vehicles With Unknown Dynamics and Unbounded Disturbances
    Bao, Hanqiu
    Kang, Qi
    Shi, Xudong
    Xiao, Lingfei
    An, Jing
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2024, 9 (02): : 3409 - 3421
  • [8] Robust discrete-time set-based adaptive predictive control for nonlinear systems
    Goncalves, Guilherme A. A.
    Guay, Martin
    JOURNAL OF PROCESS CONTROL, 2016, 39 : 111 - 122
  • [9] Robust Output Regulation and Reinforcement Learning-Based Output Tracking Design for Unknown Linear Discrete-Time Systems
    Chen, Ci
    Xie, Lihua
    Jiang, Yi
    Xie, Kan
    Xie, Shengli
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2023, 68 (04) : 2391 - 2398
  • [10] Fully Actuated System Approaches: Predictive Elimination Control for Discrete-Time Nonlinear Time-Varying Systems With Full State Constraints and Time-Varying Delays
    Wang, Xiubo
    Duan, Guangren
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2024, 71 (01) : 383 - 396