Dynamic Leader-Follower Output Containment Control of Heterogeneous Multiagent Systems Using Reinforcement Learning

被引:2
作者
Zhang, Huaipin [1 ]
Zhao, Wei [2 ]
Xie, Xiangpeng [1 ]
Yue, Dong [1 ]
机构
[1] Nanjing Univ Posts & Telecommun, Inst Adv Technol Carbon Neutral, Nanjing 210023, Peoples R China
[2] Nanjing Univ Finance & Econ, Coll Informat Engn, Nanjing 210023, Peoples R China
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2024年 / 54卷 / 09期
关键词
Observers; Vectors; Heuristic algorithms; Trajectory; Multi-agent systems; Approximation algorithms; Adaptive observers; containment control; heterogeneous multiagent systems (MASs); neural network (NN) approximation; reinforcement learning (RL);
D O I
10.1109/TSMC.2024.3406777
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This article addresses the optimal containment problem of heterogeneous multiagent systems (MASs) with dynamic leaders via reinforcement learning (RL), where the dynamics of all agents are all completely unknown. A distributed model-free observer is constructed for each follower to estimate the leaders' dynamics and the output trajectories inside the convex hull formed by the leaders. Based on the designed observers, the optimal containment problem is formulated as an optimal tracking control issue. Then the discounted performance functions are introduced to obtain algebraic Riccati equations (AREs). And a model-free RL algorithm is developed to learn the AREs online. To implement this algorithm, we design a single critic neural network structure for each follower to approximate Q -function, and estimate optimal control policy and worst-case adversarial input policy. Finally, a numerical simulation is provided to demonstrate the effectiveness of the proposed algorithm.
引用
收藏
页码:5307 / 5316
页数:10
相关论文
共 50 条
  • [1] Leader-Follower Output Synchronization of Linear Heterogeneous Systems With Active Leader Using Reinforcement Learning
    Yang, Yongliang
    Modares, Hamidreza
    Wunsch, Donald C., II
    Yin, Yixin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (06) : 2139 - 2153
  • [2] Fully Heterogeneous Containment Control of a Network of Leader-Follower Systems
    Mazouchi, Majid
    Tatari, Farzaneh
    Kiumarsi, Bahare
    Modares, Hamidreza
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (11) : 6187 - 6194
  • [3] Optimized Leader-Follower Consensus Control Using Reinforcement Learning for a Class of Second-Order Nonlinear Multiagent Systems
    Wen, Guoxing
    Li, Bin
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (09): : 5546 - 5555
  • [4] Adaptive Fuzzy Leader-Follower Synchronization of Constrained Heterogeneous Multiagent Systems
    Yang, Yongliang
    Xu, Cheng-Zhong
    IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2022, 30 (01) : 205 - 219
  • [5] Resiliency in dynamic leader-follower multiagent systems
    Rezaee, Hamed
    Parisini, Thomas
    Polycarpou, Marios M.
    AUTOMATICA, 2021, 125 (125)
  • [6] DYNAMIC EVENT-TRIGGERED LEADER-FOLLOWER CONSENSUS CONTROL FOR MULTIAGENT SYSTEMS
    Wu, Xiaoqun
    Mao, Bing
    Wu, Xiuqi
    Lu, Jinhu
    SIAM JOURNAL ON CONTROL AND OPTIMIZATION, 2022, 60 (01) : 189 - 209
  • [7] Event-Triggered Output Feedback Control for Leader-Follower Consensus of Feedforward Nonlinear Multiagent Systems
    Li, Hanfeng
    Zhang, Xianfu
    Fan, Debao
    IEEE SYSTEMS JOURNAL, 2022, 16 (04): : 6054 - 6061
  • [8] Leader-Follower Formation Learning Control of Discrete-Time Nonlinear Multiagent Systems
    Shi, Haotian
    Wang, Min
    Wang, Cong
    IEEE TRANSACTIONS ON CYBERNETICS, 2023, 53 (02) : 1184 - 1194
  • [9] Data-Driven Optimal Synchronization Control for Leader-Follower Multiagent Systems
    Zhou, Yuanqiang
    Li, Dewei
    Gao, Furong
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2023, 53 (01): : 495 - 503
  • [10] Distributed Tracking of Leader-Follower Multiagent Systems Subject to Disturbed Leader's Information
    Li, Xiangyang
    Xu, Sihan
    Gao, Huanli
    Cai, He
    IEEE ACCESS, 2020, 8 : 227970 - 227981