Model-free policy iteration optimal control of fuzzy systems via a two-player zero-sum game

被引:0
作者
Deng, Yifan [1 ]
Wu, Wei [1 ]
Tong, Shaocheng [1 ]
机构
[1] Liaoning Univ Technol, Coll Sci, Jinzhou 121001, Liaoning, Peoples R China
基金
中国国家自然科学基金;
关键词
Takagi-Sugeno (T-S) fuzzy systems; optimal control; two-player zero-sum game; policy iteration (PI) algorithm; stability and convergence; TIME-SYSTEMS; TRACKING CONTROL; STABILIZATION; DESIGN;
D O I
10.1080/00207721.2025.2480192
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we study the optimal control problem for Takagi-Sugeno (T-S) fuzzy systems with disturbances. Due to the presence of disturbances in the T-S fuzzy systems, a fuzzy optimal state feedback control approach is presented by employing a two-player zero-sum game. Since the analytical optimal control solutions and the worst-case disturbance policies of T-S fuzzy systems can be boiled down to solving the algebraic Riccati equations (AREs), which are difficult to be obtained directly, a model-free policy iteration (PI) learning algorithm is proposed to obtain their approximation solutions. It is proved that the developed fuzzy optimal state feedback controller can ensure the fuzzy systems to be asymptotically stable and satisfy the disturbance attenuation condition simultaneously. Also, the designed PI learning algorithm can converge to their optimal solutions. Finally, we apply the developed fuzzy optimal state feedback control method to the truck-trailer system and the simulation results demonstrated the effectiveness of the developed scheme.
引用
收藏
页数:13
相关论文
共 35 条
  • [1] Galerkin approximations of the generalized Hamilton-Jacobi-Bellman equation
    Beard, RW
    Saridis, GN
    Wen, JT
    [J]. AUTOMATICA, 1997, 33 (12) : 2159 - 2177
  • [2] Reinforcement Learning for Linear Continuous-time Systems: an Incremental Learning Approach
    Bian, Tao
    Jiang, Zhong-Ping
    [J]. IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2019, 6 (02) : 433 - 440
  • [3] Delay-Variation-Dependent Criteria on Stability and Stabilization for Discrete-Time T-S Fuzzy Systems With Time-Varying Delays
    Chen, Wen-Hu
    Zhang, Chuan-Ke
    Xie, Ke-You
    Zhu, Cui
    He, Yong
    [J]. IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2022, 30 (11) : 4856 - 4866
  • [4] State feedback control of continuous-time T-S fuzzy systems via switched fuzzy controllers
    Dong, Jiuxiang
    Yang, Guang-Hong
    [J]. INFORMATION SCIENCES, 2008, 178 (06) : 1680 - 1695
  • [5] Fuzzy-Based Adaptive Optimization of Unknown Discrete-Time Nonlinear Markov Jump Systems With Off-Policy Reinforcement Learning
    Fang, Haiyang
    Tu, Yidong
    Wang, Hai
    He, Shuping
    Liu, Fei
    Ding, Zhengtao
    Cheng, Shing Shin
    [J]. IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2022, 30 (12) : 5276 - 5290
  • [6] Adaptive Dynamic Programming and Adaptive Optimal Output Regulation of Linear Systems
    Gao, Weinan
    Jiang, Zhong-Ping
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2016, 61 (12) : 4164 - 4169
  • [7] Value iteration and adaptive optimal output regulation with assured convergence rate
    Jiang, Yi
    Gao, Weinan
    Na, Jing
    Zhang, Di
    Hamalainen, Timo T.
    Stojanovic, Vladimir
    Lewis, Frank L.
    [J]. CONTROL ENGINEERING PRACTICE, 2022, 121
  • [8] Robust adaptive dynamic programming for linear and nonlinear systems: An overview
    Jiang, Zhong-Ping
    Jiang, Yu
    [J]. EUROPEAN JOURNAL OF CONTROL, 2013, 19 (05) : 417 - 425
  • [9] Integral Reinforcement Learning for Linear Continuous-Time Zero-Sum Games With Completely Unknown Dynamics
    Li, Hongliang
    Liu, Derong
    Wang, Ding
    [J]. IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2014, 11 (03) : 706 - 714
  • [10] Robust Inverse Q-Learning for Continuous-Time Linear Systems in Adversarial Environments
    Lian, Bosen
    Xue, Wenqian
    Lewis, Frank L.
    Chai, Tianyou
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (12) : 13083 - 13095