Reinforcement Learning for Fuzzy Structured Adaptive Optimal Control of Discrete-Time Nonlinear Complex Networks

被引:2
作者
Wu, Tao [1 ]
Cao, Jinde [1 ,2 ,3 ]
Xiong, Lianglin [4 ]
Park, Ju H. [5 ]
Lam, Hak-Keung [6 ]
机构
[1] Southeast Univ, Sch Math, Nanjing 210096, Peoples R China
[2] Southeast Univ, Frontiers Sci Ctr Mobile Informat Commun & Secur, Nanjing 210096, Peoples R China
[3] Purple Mt Labs, Nanjing 211111, Peoples R China
[4] Yunnan Open Univ, Sch Media & Informat Engn, Kunming 650504, Peoples R China
[5] Yeungnam Univ, Dept Elect Engn, Gyongsan 38541, South Korea
[6] Kings Coll London, Dept Engn, London WC2R 2LS, England
基金
新加坡国家研究基金会;
关键词
Adaptive optimal control; discrete-time nonlinear complex networks; fuzzy coupled algebraic Riccati equations (CAREs); reinforcement learning (RL); structural learning iteration; PINNING SYNCHRONIZATION; ROBUST STABILIZATION; TRACKING CONTROL; CONTROL-SYSTEMS;
D O I
10.1109/TFUZZ.2024.3434690
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article focuses on fuzzy structural adaptive optimal control issue of discrete-time nonlinear complex networks (CNs) via adopting the reinforcement learning (RL) and Takagi-Sugeno fuzzy modeling approaches, where the control gains are subjected to structured constraints. In accordance with the Bellman optimality theory, the modified fuzzy coupled algebraic Riccati equations (CAREs) are constructed for discrete-time fuzzy CNs, while the modified fuzzy CAREs are difficult to solve directly through mathematical approaches. Then, a model-based offline learning iteration algorithm is developed to solve the modified fuzzy CAREs, where the network dynamics information is needed. Moreover, a novel data-driven off-policy RL algorithm is given to compute the modified fuzzy CAREs, and the structural optimal solutions can be obtained directly by using the collected state and input data in the absence of the network dynamics information. Furthermore, the convergence proofs of the presented learning algorithms are provided. In the end, the validity and practicability of the theoretical results are explicated via two numerical simulations.
引用
收藏
页码:6035 / 6043
页数:9
相关论文
共 44 条
[1]   Communication dynamics in complex brain networks [J].
Avena-Koenigsberger, Andrea ;
Misic, Bratislav ;
Sporns, Olaf .
NATURE REVIEWS NEUROSCIENCE, 2018, 19 (01) :17-33
[2]   Evolutionary capacitance as a general feature of complex gene networks [J].
Bergman, A ;
Siegal, ML .
NATURE, 2003, 424 (6948) :549-552
[3]   Robust exponential stabilization of stochastic coupled T-S fuzzy complex networks subject to state-dependent impulsive control [J].
Cao, Zhengran ;
Li, Chuandong ;
Zhang, Xiaoyu ;
Yang, Xujun .
INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2023, 33 (05) :3334-3357
[4]  
Furieri L, 2020, PR MACH LEARN RES, V120, P287
[5]   Fast Online Reinforcement Learning of Distributed Optimal Controller for Large-Scale Network Systems [J].
Hoshiya, Tomoki ;
Sadamoto, Tomonori .
5TH IEEE CONFERENCE ON CONTROL TECHNOLOGY AND APPLICATIONS (IEEE CCTA 2021), 2021, :1135-1141
[6]   Pinning Synchronization in T-S Fuzzy Complex Networks With Partial and Discrete-Time Couplings [J].
Huang, Chi ;
Ho, Daniel W. C. ;
Lu, Jianquan ;
Kurths, Juergen .
IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2015, 23 (04) :1274-1285
[7]  
Jain A, 2017, P AMER CONTR CONF, P2093, DOI 10.23919/ACC.2017.7963261
[8]  
Jiang Y., 2022, Automatica, V148
[9]  
Jiang Y., 2017, Robust adaptive dynamic programming
[10]   Fully distributed adaptive control for output consensus of uncertain discrete-time linear multi-agent systems [J].
Jiang, Yi ;
Liu, Lu ;
Feng, Gang .
AUTOMATICA, 2024, 162