Optimal Containment Control for Unknown Active Heterogeneous MASs via Model-Free Recursive Reinforcement Learning

被引:0
|
作者
Xia, Lina [1 ,2 ]
Li, Qing [2 ]
Song, Ruizhuo [1 ]
Yang, Gaofu [1 ]
机构
[1] Univ Sci & Technol Beijing, Beijing Engn Res Ctr Ind Spectrum Imaging, Sch Automat & Elect Engn, Beijing 100083, Peoples R China
[2] Univ Sci & Technol Beijing, Sch Automat & Elect Engn, Key Lab Knowledge Automat Ind Proc, Minist Educ, Beijing 100083, Peoples R China
来源
IEEE ACCESS | 2025年 / 13卷
基金
中国国家自然科学基金;
关键词
Observers; Protocols; Heuristic algorithms; Optimal control; Reinforcement learning; Mathematical models; Directed graphs; Convergence; Laplace equations; Convex hulls; Optimal containment control; active leaders; fully distributed observers; model-free recursive reinforcement learning; HOMOGENEOUS MULTIAGENT SYSTEMS; CONTINUOUS-TIME SYSTEMS; TRACKING CONTROL; OUTPUT REGULATION; CONSENSUS; LEADER; SYNCHRONIZATION;
D O I
10.1109/ACCESS.2025.3526871
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The distributed optimal output containment control problem for multi-agent systems (MASs) involves coordinating a group of autonomous agents to drive the outputs of all followers into the convex hull spanned by the outputs of the leaders while optimizing system performance, which has numerous applications. In this paper, a fully distributed optimal containment tracking control protocol is established for unknown active heterogeneous MASs with external disturbances. Firstly, a fully distributed observer is designed to ensure its trajectory stays within the convex hull established by active leaders without requiring global network topology information. Subsequently, an augmented system is constructed using the dynamics of the followers and the observers to design $H_{\infty }$ optimal containment control protocol. Then, a model-free recursive reinforcement learning (RRL) algorithm is devised to learn the optimal control protocol, which demonstrates that the weight iteration error asymptotically converges to zero, and the algorithm exhibits favorable convergence speed. Finally, the effectiveness of the proposed improved algorithm is validated using a heterogeneous nonlinear multi-agent model.
引用
收藏
页码:7603 / 7613
页数:11
相关论文
共 50 条
  • [31] Event-Triggered Iterative Learning Containment Control of Model-Free Multiagent Systems
    Hua, Changchun
    Qiu, Yunfei
    Guan, Xinping
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2021, 51 (12): : 7719 - 7726
  • [32] Model-free reinforcement learning approach to optimal speed control of combustion engines in start-up mode
    Xu, Zhenhui
    Pan, Linjun
    Shen, Tielong
    CONTROL ENGINEERING PRACTICE, 2021, 111
  • [33] Optimal consensus control for unknown second-order multi-agent systems: Using model-free reinforcement learning method
    Li, Jun
    Ji, Lianghao
    Li, Huaqing
    APPLIED MATHEMATICS AND COMPUTATION, 2021, 410
  • [34] Optimal model-free adaptive control based on reinforcement Q-Learning for solar thermal collector fields
    Pataro, Igor M. L.
    Cunha, Rita
    Gil, Juan D.
    Guzman, Jose L.
    Berenguel, Manuel
    Lemos, Joao M.
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 126
  • [35] Dynamic Event-Triggered Model-Free Reinforcement Learning for Cooperative Control of Multiagent Systems
    Wang, Ke
    Tang, Zhuo
    Mu, Chaoxu
    IEEE TRANSACTIONS ON RELIABILITY, 2024,
  • [36] Model-free optimal containment control of multi-agent systems based on actor-critic framework
    Wang, Wei
    Chen, Xin
    NEUROCOMPUTING, 2018, 314 : 242 - 250
  • [37] Iterative Q-Learning for Model-Free Optimal Control With Adjustable Convergence Rate
    Wang, Ding
    Wang, Yuan
    Zhao, Mingming
    Qiao, Junfei
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2024, 71 (04) : 2224 - 2228
  • [38] On Distributed Model-Free Reinforcement Learning Control With Stability Guarantee
    Mukherjee, Sayak
    Vu, Thanh Long
    IEEE CONTROL SYSTEMS LETTERS, 2021, 5 (05): : 1615 - 1620
  • [39] Model-Free Reinforcement Learning of Impedance Control in Stochastic Environments
    Stulp, Freek
    Buchli, Jonas
    Ellmer, Alice
    Mistry, Michael
    Theodorou, Evangelos A.
    Schaal, Stefan
    IEEE TRANSACTIONS ON AUTONOMOUS MENTAL DEVELOPMENT, 2012, 4 (04) : 330 - 341
  • [40] Optimal bipartite consensus control for heterogeneous unknown multi-agent systems via reinforcement learning
    Meng, Hao
    Pang, Denghao
    Cao, Jinde
    Guo, Yechen
    Niazi, Azmat Ullah Khan
    APPLIED MATHEMATICS AND COMPUTATION, 2024, 476