Robust Learning-Based Predictive Control for Discrete-Time Nonlinear Systems With Unknown Dynamics and State Constraints

被引:33
|
作者
Zhang, Xinglong [1 ]
Liu, Jiahang [1 ]
Xu, Xin [1 ]
Yu, Shuyou [2 ,3 ]
Chen, Hong [4 ]
机构
[1] Natl Univ Def Technol, Coll Intelligence Sci & Technol, Changsha 410073, Peoples R China
[2] Jilin Univ, State Key Lab Automot Simulat & Control, Changchun 130025, Peoples R China
[3] Jilin Univ, Dept Control Sci & Engn, Changchun 130025, Peoples R China
[4] Tongji Univ, Coll Elect & Informat Engn, Shanghai 201804, Peoples R China
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2022年 / 52卷 / 12期
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Robustness; Predictive control; Aerospace electronics; Computational modeling; Predictive models; Optimal control; Heuristic algorithms; Model predictive control (MPC); nonlinear systems; reinforcement learning (RL); robustness; state constraints; TRACKING CONTROL; LINEAR-SYSTEMS; STABILITY;
D O I
10.1109/TSMC.2022.3146284
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Robust model predictive control (MPC) is a well-known control technique for model-based control with constraints and uncertainties. In classic robust tube-based MPC approaches, an open-loop control sequence is computed via periodically solving an online nominal MPC problem, which requires prior model information and frequent access to onboard computational resources. In this article, we propose an efficient robust MPC solution based on receding horizon reinforcement learning, called r-LPC, for unknown nonlinear systems with state constraints and disturbances. The proposed r-LPC utilizes a Koopman operator-based prediction model obtained offline from precollected input-output datasets. Unlike classic tube-based MPC, in each prediction time interval of r-LPC, we use an actor-critic structure to learn a near-optimal feedback control policy rather than a control sequence. The resulting closed-loop control policy can be learned offline and deployed online or learned online in an asynchronous way. In the latter case, online learning can be activated whenever necessary; for instance, the safety constraint is violated with the deployed policy. The closed-loop recursive feasibility, robustness, and asymptotic stability are proven under function approximation errors of the actor-critic networks. Simulation and experimental results on two nonlinear systems with unknown dynamics and disturbances have demonstrated that our approach has better or comparable performance when compared with tube-based MPC and linear quadratic regulator, and outperforms a recently developed actor-critic learning approach.
引用
收藏
页码:7314 / 7327
页数:14
相关论文
共 50 条
  • [31] Stabilization of Discrete-time Switched Systems with State Constraints Based on Mode-Dependent Average Dwell Time
    Su, Qingyu
    Wang, Peipei
    Li, Jian
    Liu, Honghai
    ASIAN JOURNAL OF CONTROL, 2017, 19 (01) : 67 - 73
  • [32] Optimal Output Regulation of Linear Discrete-Time Systems With Unknown Dynamics Using Reinforcement Learning
    Jiang, Yi
    Kiumarsi, Bahare
    Fan, Jialu
    Chai, Tianyou
    Li, Jinna
    Lewis, Frank L.
    IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (07) : 3147 - 3156
  • [33] Robust adaptive output feedback control of a class of discrete-time nonlinear systems with nonlinear uncertainties and unknown control directions
    Dai, Shi-Lu
    Yang, Chenguang
    Ge, Shuzhi Sam
    Lee, Tong Heng
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2013, 23 (13) : 1472 - 1495
  • [34] Robust Fuzzy Predictive Control for Discrete-Time Systems With Interval Time-Varying Delays and Unknown Disturbances
    Shi, Huiyuan
    Li, Ping
    Cao, Jiangtao
    Su, Chengli
    Yu, Jingxian
    IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2020, 28 (07) : 1504 - 1516
  • [35] Reference regulation of constrained uncertain discrete-time systems via robust predictive integral control
    Ghaffari, Valiollah
    INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2024, 55 (01) : 1 - 15
  • [36] NN-Based Reinforcement Learning Optimal Control for Inequality-Constrained Nonlinear Discrete-Time Systems With Disturbances
    Li, Shu
    Ding, Liang
    Zheng, Miao
    Liu, Zixuan
    Li, Xinyu
    Yang, Huaiguang
    Gao, Haibo
    Deng, Zongquan
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (11) : 15507 - 15516
  • [37] On Quadratic Optimal Control of Nonlinear Discrete-Time Systems
    Elloumi, Salwa
    Mechichi, Amina Khiari
    Braiek, Naceur Benhadj
    2013 10TH INTERNATIONAL MULTI-CONFERENCE ON SYSTEMS, SIGNALS & DEVICES (SSD), 2013,
  • [38] Stabilization of uncertain nonlinear discrete-time switched systems with state delays: A constrained robust model predictive control approach
    Aminsafaee, Maryam
    Shafiei, Mohammad Hossein
    JOURNAL OF VIBRATION AND CONTROL, 2019, 25 (14) : 2079 - 2090
  • [39] Reinforcement Learning-Based Adaptive Optimal Exponential Tracking Control of Linear Systems With Unknown Dynamics
    Chen, Ci
    Modares, Hamidreza
    Xie, Kan
    Lewis, Frank L.
    Wan, Yan
    Xie, Shengli
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2019, 64 (11) : 4423 - 4438
  • [40] Unknown inputs observers for state and unknown inputs estimation in a class of discrete-time Lipschitz nonlinear systems
    Seifeddine, Ben Warrad
    Slim, Dhahri
    Faycal, Ben Hmida
    2013 INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING AND SOFTWARE APPLICATIONS (ICEESA), 2013, : 157 - 161