Reinforcement Learning for Fuzzy Structured Adaptive Optimal Control of Discrete-Time Nonlinear Complex Networks

被引:2
作者
Wu, Tao [1 ]
Cao, Jinde [1 ,2 ,3 ]
Xiong, Lianglin [4 ]
Park, Ju H. [5 ]
Lam, Hak-Keung [6 ]
机构
[1] Southeast Univ, Sch Math, Nanjing 210096, Peoples R China
[2] Southeast Univ, Frontiers Sci Ctr Mobile Informat Commun & Secur, Nanjing 210096, Peoples R China
[3] Purple Mt Labs, Nanjing 211111, Peoples R China
[4] Yunnan Open Univ, Sch Media & Informat Engn, Kunming 650504, Peoples R China
[5] Yeungnam Univ, Dept Elect Engn, Gyongsan 38541, South Korea
[6] Kings Coll London, Dept Engn, London WC2R 2LS, England
基金
新加坡国家研究基金会;
关键词
Adaptive optimal control; discrete-time nonlinear complex networks; fuzzy coupled algebraic Riccati equations (CAREs); reinforcement learning (RL); structural learning iteration; PINNING SYNCHRONIZATION; ROBUST STABILIZATION; TRACKING CONTROL; CONTROL-SYSTEMS;
D O I
10.1109/TFUZZ.2024.3434690
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article focuses on fuzzy structural adaptive optimal control issue of discrete-time nonlinear complex networks (CNs) via adopting the reinforcement learning (RL) and Takagi-Sugeno fuzzy modeling approaches, where the control gains are subjected to structured constraints. In accordance with the Bellman optimality theory, the modified fuzzy coupled algebraic Riccati equations (CAREs) are constructed for discrete-time fuzzy CNs, while the modified fuzzy CAREs are difficult to solve directly through mathematical approaches. Then, a model-based offline learning iteration algorithm is developed to solve the modified fuzzy CAREs, where the network dynamics information is needed. Moreover, a novel data-driven off-policy RL algorithm is given to compute the modified fuzzy CAREs, and the structural optimal solutions can be obtained directly by using the collected state and input data in the absence of the network dynamics information. Furthermore, the convergence proofs of the presented learning algorithms are provided. In the end, the validity and practicability of the theoretical results are explicated via two numerical simulations.
引用
收藏
页码:6035 / 6043
页数:9
相关论文
共 50 条
[21]   Q-learning solution for optimal consensus control of discrete-time multiagent systems using reinforcement learning [J].
Mu, Chaoxu ;
Zhao, Qian ;
Gao, Zhongke ;
Sun, Changyin .
JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2019, 356 (13) :6946-6967
[22]   H∞ Static Output-Feedback Control Design for Discrete-Time Systems Using Reinforcement Learning [J].
Valadbeigi, Amir Parviz ;
Sedigh, Ali Khaki ;
Lewis, F. L. .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (02) :396-406
[23]   Nonlinear discrete time optimal control based on Fuzzy Models [J].
Jin, Xin ;
Shin, Yung C. .
JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2015, 29 (02) :647-658
[24]   Adaptive Reinforcement Learning Control Based on Neural Approximation for Nonlinear Discrete-Time Systems With Unknown Nonaffine Dead-Zone Input [J].
Liu, Yan-Jun ;
Li, Shu ;
Tong, Shaocheng ;
Chen, C. L. Philip .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (01) :295-305
[25]   A Novel Adaptive Fuzzy Control for a Class of Discrete-Time Nonlinear Systems in Strict-Feedback Form [J].
Wang, Xin ;
Li, Tieshan ;
Lin, Bin .
2014 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS (FUZZ-IEEE), 2014, :727-733
[26]   Infinite Horizon Self-Learning Optimal Control of Nonaffine Discrete-Time Nonlinear Systems [J].
Wei, Qinglai ;
Liu, Derong ;
Yang, Xiong .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (04) :866-879
[27]   Reinforcement Q-learning for optimal tracking control of linear discrete-time systems with unknown dynamics [J].
Kiumarsi, Bahare ;
Lewis, Frank L. ;
Modares, Hamidreza ;
Karimpour, Ali ;
Naghibi-Sistani, Mohammad-Bagher .
AUTOMATICA, 2014, 50 (04) :1167-1175
[28]   Direct adaptive fuzzy backstepping control for uncertain discrete-time nonlinear systems using noisy measurements [J].
Yoshimura, Toshio .
INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2017, 48 (04) :695-704
[29]   Optimal Output Regulation of Linear Discrete-Time Systems With Unknown Dynamics Using Reinforcement Learning [J].
Jiang, Yi ;
Kiumarsi, Bahare ;
Fan, Jialu ;
Chai, Tianyou ;
Li, Jinna ;
Lewis, Frank L. .
IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (07) :3147-3156
[30]   Reinforcement Learning Based Adaptive Predefined-Time Optimal Control for Strict-Feedback Nonlinear Systems [J].
Jin, Yitong ;
Wang, Fang ;
Zhang, Xueyi .
INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2025,