Optimized Backstepping Consensus Control Using Reinforcement Learning for a Class of Nonlinear Strict-Feedback-Dynamic Multi-Agent Systems

被引:67
|
作者
Wen, Guoxing [1 ,2 ]
Chen, C. L. Philip [3 ,4 ]
机构
[1] Binzhou Univ, Coll Sci, Binzhou 256600, Peoples R China
[2] Qilu Univ Technol, Sch Math & Stat, Jinan 250353, Peoples R China
[3] South China Univ Technol, Sch Comp Sci & Engn, Guangzhou 510641, Peoples R China
[4] Dalian Maritime Univ, Nav Coll, Dalian 116026, Peoples R China
基金
中国国家自然科学基金;
关键词
Optimal control; Backstepping; Artificial neural networks; Performance analysis; Nonlinear dynamical systems; Consensus control; Mathematical model; Critic-actor architecture; high-order multi-agent system (MAS); neural network (NN); optimal control; reinforcement learning (RL); ADAPTIVE OPTIMAL-CONTROL; CONTAINMENT CONTROL; TRACKING CONTROL;
D O I
10.1109/TNNLS.2021.3105548
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this article, an optimized leader-following consensus control scheme is proposed for the nonlinear strict-feedback-dynamic multi-agent system by learning from the controlling idea of optimized backstepping technique, which designs the virtual and actual controls of backstepping to be the optimized solution of corresponding subsystems so that the entire backstepping control is optimized. Since this control needs to not only ensure the optimizing system performance but also synchronize the multiple system state variables, it is an interesting and challenging topic. In order to achieve this optimized control, the neural network approximation-based reinforcement learning (RL) is performed under critic-actor architecture. In most of the existing RL-based optimal controls, since both the critic and actor RL updating laws are derived from the negative gradient of square of the Hamilton-Jacobi-Bellman (HJB) equation's approximation, which contains multiple nonlinear terms, their algorithm are inevitably intricate. However, the proposed optimized control derives the RL updating laws from the negative gradient of a simple positive function, which is correlated with the HJB equation; hence, it can be significantly simple in the algorithm. Meanwhile, it can also release two general conditions, known dynamic and persistence excitation, which are required in most of the RL-based optimal controls. Therefore, the proposed optimized scheme can be a natural selection for the high-order nonlinear multi-agent control. Finally, the effectiveness is demonstrated by both theory and simulation.
引用
收藏
页码:1524 / 1536
页数:13
相关论文
共 50 条
  • [21] Adaptive Neural Output Consensus Control of Stochastic Nonlinear Strict-Feedback Multi-Agent Systems
    Yang, Yang
    Miao, Songtao
    Xu, Chuang
    Yue, Dong
    Tan, Jie
    Tian, Yu-Chu
    2018 AUSTRALIAN & NEW ZEALAND CONTROL CONFERENCE (ANZCC), 2018, : 385 - 390
  • [22] Adaptive Neural Network Optimized Control Using Reinforcement Learning of Critic-Actor Architecture for a Class of Non-Affine Nonlinear Systems
    Yang, Xue
    Li, Bin
    Wen, Guoxing
    IEEE ACCESS, 2021, 9 : 141758 - 141765
  • [23] Adaptive containment control for nonlinear strict-feedback multi-agent systems with dynamic leaders
    Cui, Yang
    Liu, Xiaoping
    Deng, Xin
    Wang, Lidong
    INTERNATIONAL JOURNAL OF CONTROL, 2022, 95 (06) : 1658 - 1667
  • [24] Optimized Backstepping Combined With Dynamic Surface Technique for Single-Input-Single-Output Nonlinear Strict-Feedback System
    Wen, Guoxing
    Zhou, Ranran
    Zhao, Yanlong
    Niu, Ben
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (07): : 4210 - 4221
  • [25] Deterministic Reinforcement Learning Consensus Control of Nonlinear Multi-Agent Systems via Autonomous Convergence Perception
    Gao, Shigen
    Xu, Chaoan
    Dong, Hairong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2024, 71 (04) : 2229 - 2233
  • [26] Reinforcement Learning Control for Consensus of the Leader-Follower Multi-Agent Systems
    Chiang, Ming-Li
    Liu, An-Sheng
    Fu, Li-Chen
    PROCEEDINGS OF 2018 IEEE 7TH DATA DRIVEN CONTROL AND LEARNING SYSTEMS CONFERENCE (DDCLS), 2018, : 1152 - 1157
  • [27] Optimized tracking control using reinforcement learning strategy for a class of nonlinear systems
    Yang, Xue
    Li, Bin
    ASIAN JOURNAL OF CONTROL, 2023, 25 (03) : 2095 - 2104
  • [28] Optimized Inverse Dead-Zone Control Using Reinforcement Learning for a Class of Nonlinear Systems
    Sun, Wenxia
    Ma, Shuaihua
    Li, Bin
    Wen, Guoxing
    INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2024, 38 (12) : 3855 - 3864
  • [29] Optimized Formation Control Using Simplified Reinforcement Learning for a Class of Multiagent Systems With Unknown Dynamics
    Wen, Guoxing
    Chen, C. L. Philip
    Li, Bin
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2020, 67 (09) : 7879 - 7888
  • [30] Optimized tracking control using reinforcement learning and backstepping technique for canonical nonlinear unknown dynamic system
    Song, Yanfen
    Li, Zijun
    Wen, Guoxing
    OPTIMAL CONTROL APPLICATIONS & METHODS, 2024, 45 (04) : 1655 - 1671