Optimized Backstepping Consensus Control Using Reinforcement Learning for a Class of Nonlinear Strict-Feedback-Dynamic Multi-Agent Systems

被引:67
作者
Wen, Guoxing [1 ,2 ]
Chen, C. L. Philip [3 ,4 ]
机构
[1] Binzhou Univ, Coll Sci, Binzhou 256600, Peoples R China
[2] Qilu Univ Technol, Sch Math & Stat, Jinan 250353, Peoples R China
[3] South China Univ Technol, Sch Comp Sci & Engn, Guangzhou 510641, Peoples R China
[4] Dalian Maritime Univ, Nav Coll, Dalian 116026, Peoples R China
基金
中国国家自然科学基金;
关键词
Optimal control; Backstepping; Artificial neural networks; Performance analysis; Nonlinear dynamical systems; Consensus control; Mathematical model; Critic-actor architecture; high-order multi-agent system (MAS); neural network (NN); optimal control; reinforcement learning (RL); ADAPTIVE OPTIMAL-CONTROL; CONTAINMENT CONTROL; TRACKING CONTROL;
D O I
10.1109/TNNLS.2021.3105548
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this article, an optimized leader-following consensus control scheme is proposed for the nonlinear strict-feedback-dynamic multi-agent system by learning from the controlling idea of optimized backstepping technique, which designs the virtual and actual controls of backstepping to be the optimized solution of corresponding subsystems so that the entire backstepping control is optimized. Since this control needs to not only ensure the optimizing system performance but also synchronize the multiple system state variables, it is an interesting and challenging topic. In order to achieve this optimized control, the neural network approximation-based reinforcement learning (RL) is performed under critic-actor architecture. In most of the existing RL-based optimal controls, since both the critic and actor RL updating laws are derived from the negative gradient of square of the Hamilton-Jacobi-Bellman (HJB) equation's approximation, which contains multiple nonlinear terms, their algorithm are inevitably intricate. However, the proposed optimized control derives the RL updating laws from the negative gradient of a simple positive function, which is correlated with the HJB equation; hence, it can be significantly simple in the algorithm. Meanwhile, it can also release two general conditions, known dynamic and persistence excitation, which are required in most of the RL-based optimal controls. Therefore, the proposed optimized scheme can be a natural selection for the high-order nonlinear multi-agent control. Finally, the effectiveness is demonstrated by both theory and simulation.
引用
收藏
页码:1524 / 1536
页数:13
相关论文
共 50 条
  • [31] Optimized inverse dead-zone formation control using reinforcement learning for the nonlinear single-integrator dynamic multi-agent system
    Wen, Guoxing
    Sun, Wenxia
    Ma, Shuaihua
    NEUROCOMPUTING, 2025, 636
  • [32] Command-filter-based adaptive finite-time consensus control for nonlinear strict-feedback multi-agent systems with dynamic leader
    Cui, Yang
    Liu, Xiaoping
    Deng, Xin
    Wen, Guoxing
    INFORMATION SCIENCES, 2021, 565 : 17 - 31
  • [33] Optimized Backstepping Tracking Control Using Reinforcement Learning for Quadrotor Unmanned Aerial Vehicle System
    Wen, Guoxing
    Hao, Wei
    Feng, Weiwei
    Gao, Kaizhou
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (08): : 5004 - 5015
  • [34] Distributed Consensus Control Using Neural Network for a Class of Nonlinear Multi-agent Systems
    Wen, Guo-Xing
    Chen, C. L. Philip
    PROCEEDINGS 2012 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2012, : 2591 - 2595
  • [35] Simplified optimized control using reinforcement learning algorithm for a class of stochastic nonlinear systems
    Wen, Guoxing
    Chen, C. L. Philip
    Li, Wei Nian
    INFORMATION SCIENCES, 2020, 517 : 230 - 243
  • [36] Event-triggered H∞ consensus control for input-constrained multi-agent systems via reinforcement learning
    Zhang, Jinxuan
    Ren, Chang-E
    CONTROL THEORY AND TECHNOLOGY, 2024, 22 (01) : 25 - 38
  • [37] Output Feedback-Based Neural Adaptive Finite-Time Containment Control of Non-Strict Feedback Nonlinear Multi-Agent Systems
    Zhao, Lin
    Chen, Xiao
    Yu, Jinpeng
    Shi, Peng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2022, 69 (02) : 847 - 858
  • [38] Distributed adaptive consensus control of Lipschitz nonlinear multi-agent systems using output feedback
    Jameel, Atif
    Rehan, Muhammad
    Hong, Keum-Shik
    Iqbal, Naeem
    INTERNATIONAL JOURNAL OF CONTROL, 2016, 89 (11) : 2336 - 2349
  • [39] Distributed Output-Feedback Tracking Control for a Class of Nonlinear Multi-Agent Systems With Polynomial Growth
    Liu, Jinxiu
    Wang, Meiqiao
    Wang, Hui
    Li, Wuquan
    Liu, Fang
    IEEE ACCESS, 2024, 12 : 160245 - 160253
  • [40] Adaptive fuzzy backstepping control for a class of uncertain nonlinear strict-feedback systems based on dynamic surface control approach
    Peng, Jinzhu
    Dubay, Rickey
    EXPERT SYSTEMS WITH APPLICATIONS, 2019, 120 : 239 - 252