Reinforcement Learning for Fuzzy Structured Adaptive Optimal Control of Discrete-Time Nonlinear Complex Networks

被引:1
作者
Wu, Tao [1 ]
Cao, Jinde [1 ,2 ,3 ]
Xiong, Lianglin [4 ]
Park, Ju H. [5 ]
Lam, Hak-Keung [6 ]
机构
[1] Southeast Univ, Sch Math, Nanjing 210096, Peoples R China
[2] Southeast Univ, Frontiers Sci Ctr Mobile Informat Commun & Secur, Nanjing 210096, Peoples R China
[3] Purple Mt Labs, Nanjing 211111, Peoples R China
[4] Yunnan Open Univ, Sch Media & Informat Engn, Kunming 650504, Peoples R China
[5] Yeungnam Univ, Dept Elect Engn, Gyongsan 38541, South Korea
[6] Kings Coll London, Dept Engn, London WC2R 2LS, England
基金
新加坡国家研究基金会;
关键词
Adaptive optimal control; discrete-time nonlinear complex networks; fuzzy coupled algebraic Riccati equations (CAREs); reinforcement learning (RL); structural learning iteration; PINNING SYNCHRONIZATION; ROBUST STABILIZATION; TRACKING CONTROL; CONTROL-SYSTEMS;
D O I
10.1109/TFUZZ.2024.3434690
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article focuses on fuzzy structural adaptive optimal control issue of discrete-time nonlinear complex networks (CNs) via adopting the reinforcement learning (RL) and Takagi-Sugeno fuzzy modeling approaches, where the control gains are subjected to structured constraints. In accordance with the Bellman optimality theory, the modified fuzzy coupled algebraic Riccati equations (CAREs) are constructed for discrete-time fuzzy CNs, while the modified fuzzy CAREs are difficult to solve directly through mathematical approaches. Then, a model-based offline learning iteration algorithm is developed to solve the modified fuzzy CAREs, where the network dynamics information is needed. Moreover, a novel data-driven off-policy RL algorithm is given to compute the modified fuzzy CAREs, and the structural optimal solutions can be obtained directly by using the collected state and input data in the absence of the network dynamics information. Furthermore, the convergence proofs of the presented learning algorithms are provided. In the end, the validity and practicability of the theoretical results are explicated via two numerical simulations.
引用
收藏
页码:6035 / 6043
页数:9
相关论文
共 50 条
  • [21] H∞ Static Output-Feedback Control Design for Discrete-Time Systems Using Reinforcement Learning
    Valadbeigi, Amir Parviz
    Sedigh, Ali Khaki
    Lewis, F. L.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (02) : 396 - 406
  • [22] Nonlinear discrete time optimal control based on Fuzzy Models
    Jin, Xin
    Shin, Yung C.
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2015, 29 (02) : 647 - 658
  • [23] Adaptive Reinforcement Learning Control Based on Neural Approximation for Nonlinear Discrete-Time Systems With Unknown Nonaffine Dead-Zone Input
    Liu, Yan-Jun
    Li, Shu
    Tong, Shaocheng
    Chen, C. L. Philip
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (01) : 295 - 305
  • [24] Reinforcement Q-learning for optimal tracking control of linear discrete-time systems with unknown dynamics
    Kiumarsi, Bahare
    Lewis, Frank L.
    Modares, Hamidreza
    Karimpour, Ali
    Naghibi-Sistani, Mohammad-Bagher
    AUTOMATICA, 2014, 50 (04) : 1167 - 1175
  • [25] Infinite Horizon Self-Learning Optimal Control of Nonaffine Discrete-Time Nonlinear Systems
    Wei, Qinglai
    Liu, Derong
    Yang, Xiong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (04) : 866 - 879
  • [26] A Novel Adaptive Fuzzy Control for a Class of Discrete-Time Nonlinear Systems in Strict-Feedback Form
    Wang, Xin
    Li, Tieshan
    Lin, Bin
    2014 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS (FUZZ-IEEE), 2014, : 727 - 733
  • [27] Optimal Output Regulation of Linear Discrete-Time Systems With Unknown Dynamics Using Reinforcement Learning
    Jiang, Yi
    Kiumarsi, Bahare
    Fan, Jialu
    Chai, Tianyou
    Li, Jinna
    Lewis, Frank L.
    IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (07) : 3147 - 3156
  • [28] Direct adaptive fuzzy backstepping control for uncertain discrete-time nonlinear systems using noisy measurements
    Yoshimura, Toshio
    INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2017, 48 (04) : 695 - 704
  • [29] Event-Triggered Optimal Parallel Tracking Control for Discrete-Time Nonlinear Systems
    Lu, Jingwei
    Wei, Qinglai
    Liu, Yujia
    Zhou, Tianmin
    Wang, Fei-Yue
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (06): : 3772 - 3784
  • [30] Reinforcement learning-based online adaptive controller design for a class of unknown nonlinear discrete-time systems with time delays
    Liang, Yuling
    Zhang, Huaguang
    Xiao, Geyang
    Jiang, He
    NEURAL COMPUTING & APPLICATIONS, 2018, 30 (06) : 1733 - 1745