Reinforcement Learning for Fuzzy Structured Adaptive Optimal Control of Discrete-Time Nonlinear Complex Networks

被引:2
作者
Wu, Tao [1 ]
Cao, Jinde [1 ,2 ,3 ]
Xiong, Lianglin [4 ]
Park, Ju H. [5 ]
Lam, Hak-Keung [6 ]
机构
[1] Southeast Univ, Sch Math, Nanjing 210096, Peoples R China
[2] Southeast Univ, Frontiers Sci Ctr Mobile Informat Commun & Secur, Nanjing 210096, Peoples R China
[3] Purple Mt Labs, Nanjing 211111, Peoples R China
[4] Yunnan Open Univ, Sch Media & Informat Engn, Kunming 650504, Peoples R China
[5] Yeungnam Univ, Dept Elect Engn, Gyongsan 38541, South Korea
[6] Kings Coll London, Dept Engn, London WC2R 2LS, England
基金
新加坡国家研究基金会;
关键词
Adaptive optimal control; discrete-time nonlinear complex networks; fuzzy coupled algebraic Riccati equations (CAREs); reinforcement learning (RL); structural learning iteration; PINNING SYNCHRONIZATION; ROBUST STABILIZATION; TRACKING CONTROL; CONTROL-SYSTEMS;
D O I
10.1109/TFUZZ.2024.3434690
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article focuses on fuzzy structural adaptive optimal control issue of discrete-time nonlinear complex networks (CNs) via adopting the reinforcement learning (RL) and Takagi-Sugeno fuzzy modeling approaches, where the control gains are subjected to structured constraints. In accordance with the Bellman optimality theory, the modified fuzzy coupled algebraic Riccati equations (CAREs) are constructed for discrete-time fuzzy CNs, while the modified fuzzy CAREs are difficult to solve directly through mathematical approaches. Then, a model-based offline learning iteration algorithm is developed to solve the modified fuzzy CAREs, where the network dynamics information is needed. Moreover, a novel data-driven off-policy RL algorithm is given to compute the modified fuzzy CAREs, and the structural optimal solutions can be obtained directly by using the collected state and input data in the absence of the network dynamics information. Furthermore, the convergence proofs of the presented learning algorithms are provided. In the end, the validity and practicability of the theoretical results are explicated via two numerical simulations.
引用
收藏
页码:6035 / 6043
页数:9
相关论文
共 50 条
[41]   Adaptive control of a class of switched nonlinear discrete-time systems with unknown parameter [J].
Wang, Hao ;
Liu, Yan-Jun ;
Tong, Shaocheng .
NEUROCOMPUTING, 2016, 214 :1-6
[42]   Adaptive State Safety Control Design for Uncertain Discrete-Time Nonlinear Systems [J].
Zhang, Yanqi ;
Wang, Xin ;
Wang, Zhenlei .
IEEE CONTROL SYSTEMS LETTERS, 2022, 6 :2246-2251
[43]   Prescribed-Time Adaptive Fuzzy Optimal Control for Nonlinear Systems [J].
Zhang, Yan ;
Chadli, Mohammed ;
Xiang, Zhengrong .
IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2024, 32 (04) :2403-2412
[44]   Distributed Optimal Consensus Problem of Input Constrained Nonlinear Discrete-Time MASs: A Mode-Free Reinforcement Learning Approach [J].
Xuan, Shuxing ;
Liang, Hongjing ;
Huang, Shihao ;
Li, Tieshan ;
Sun, Jiayue .
IEEE TRANSACTIONS ON CYBERNETICS, 2025, 55 (06) :2910-2923
[45]   Off-Policy Reinforcement Learning for Optimal Preview Tracking Control of Linear Discrete-Time systems with unknown dynamics [J].
Wang, Chao-Ran ;
Wu, Huai-Ning .
2018 CHINESE AUTOMATION CONGRESS (CAC), 2018, :1402-1407
[47]   Adaptive fuzzy logic system-based fault tolerant control for a class of discrete-time switched nonlinear systems [J].
Li, Guangshi .
ADVANCES IN MECHANICAL ENGINEERING, 2019, 11 (05)
[48]   Off-Policy Interleaved Q-Learning: Optimal Control for Affine Nonlinear Discrete-Time Systems [J].
Li, Jinna ;
Chai, Tianyou ;
Lewis, Frank L. ;
Ding, Zhengtao ;
Jiang, Yi .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (05) :1308-1320
[49]   Adaptive Discrete-Time Flight Control Using Disturbance Observer and Neural Networks [J].
Shao, Shuyi ;
Chen, Mou ;
Zhang, Youmin .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (12) :3708-3721
[50]   Adaptive neural event-triggered near-optimal control for affined uncertain nonlinear discrete-time system [J].
Li, Xinyu ;
Ding, Liang ;
Li, Shu ;
Yang, Huaiguang ;
Qi, Huanan ;
Gao, Haibo ;
Deng, Zongquan .
ASIAN JOURNAL OF CONTROL, 2024, 26 (06) :3210-3225