Reinforcement Learning for Synchronization of Heterogeneous Multiagent Systems by Improved Q-Functions

被引:0
|
作者
Li, Jinna [1 ]
Yuan, Lin [1 ]
Cheng, Weiran [1 ]
Chai, Tianyou [2 ]
Lewis, Frank L. [3 ]
机构
[1] Liaoning Petrochem Univ, Sch Informat & Control Engn, Fushun 113001, Liaoning, Peoples R China
[2] Northeastern Univ, State Key Lab Synthet Automat Proc Ind, Shenyang 110819, Peoples R China
[3] Univ Texas Arlington, UTA Res Inst, Arlington, TX 76118 USA
基金
中国国家自然科学基金;
关键词
Synchronization; Protocols; Heuristic algorithms; Decision making; Nash equilibrium; Multi-agent systems; Games; Data-driven control; distributed control; multiagent systems (MASs); reinforcement learning (RL); synchronization;
D O I
10.1109/TCYB.2024.3440333
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This article dedicates to investigating a methodology for enhancing adaptability to environmental changes of reinforcement learning (RL) techniques with data efficiency, by which a joint control protocol is learned using only data for multiagent systems (MASs). Thus, all followers are able to synchronize themselves with the leader and minimize their individual performance. To this end, an optimal synchronization problem of heterogeneous MASs is first formulated, and then an arbitration RL mechanism is developed for well addressing key challenges faced by the current RL techniques, that is, insufficient data and environmental changes. In the developed mechanism, an improved Q-function with an arbitration factor is designed for accommodating the fact that control protocols tend to be made by historic experiences and instinctive decision-making, such that the degree of control over agents' behaviors can be adaptively allocated by on-policy and off-policy RL techniques for the optimal multiagent synchronization problem. Finally, an arbitration RL algorithm with critic-only neural networks is proposed, and theoretical analysis and proofs of synchronization and performance optimality are provided. Simulation results verify the effectiveness of the proposed method.
引用
收藏
页码:6545 / 6558
页数:14
相关论文
共 50 条
  • [1] Data-Based Optimal Synchronization of Heterogeneous Multiagent Systems in Graphical Games via Reinforcement Learning
    Xiong, Chunping
    Ma, Qian
    Guo, Jian
    Lewis, Frank L.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (11) : 15984 - 15992
  • [2] Adaptive Autonomous Synchronization of a Class of Heterogeneous Multiagent Systems
    Chen, Zhiyong
    Yan, Yamin
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2025, 70 (03) : 2066 - 2073
  • [3] Optimal Group Consensus of Multiagent Systems in Graphical Games Using Reinforcement Learning
    Wang, Yuhan
    Wang, Zhuping
    Zhang, Hao
    Yan, Huaicheng
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2025, 55 (03): : 2343 - 2353
  • [4] Group Formation Tracking of Heterogeneous Multiagent Systems Using Reinforcement Learning
    Wang, Yuhan
    Wang, Zhuping
    Zhang, Hao
    Yan, Huaicheng
    IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2025, 12 (01): : 497 - 509
  • [5] Adaptive Output Synchronization With Designated Convergence Rate of Multiagent Systems Based on Off-Policy Reinforcement Learning
    Huang, Chengjie
    Chen, Ci
    Xie, Kan
    Li, Zhenni
    Xie, Shengli
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (08): : 4667 - 4678
  • [6] Observer-Based Human-in-the-Loop Optimal Output Cluster Synchronization Control for Multiagent Systems: A Model-Free Reinforcement Learning Method
    Huang, Zongsheng
    Li, Tieshan
    Long, Yue
    Liang, Hongjing
    IEEE TRANSACTIONS ON CYBERNETICS, 2025, 55 (02) : 649 - 660
  • [7] Off-Policy Reinforcement Learning for Synchronization in Multiagent Graphical Games
    Li, Jinna
    Modares, Hamidreza
    Chai, Tianyou
    Lewis, Frank L.
    Xie, Lihua
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (10) : 2434 - 2445
  • [8] Data-Driven H∞ Output Consensus for Heterogeneous Multiagent Systems Under Switching Topology via Reinforcement Learning
    Liu, Qiwei
    Yan, Huaicheng
    Zhang, Hao
    Wang, Meng
    Tian, Yongxiao
    IEEE TRANSACTIONS ON CYBERNETICS, 2024, : 7865 - 7876
  • [9] Data-Based Optimal Consensus Control for Multiagent Systems With Policy Gradient Reinforcement Learning
    Yang, Xindi
    Zhang, Hao
    Wang, Zhuping
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (08) : 3872 - 3883
  • [10] Dynamic Leader-Follower Output Containment Control of Heterogeneous Multiagent Systems Using Reinforcement Learning
    Zhang, Huaipin
    Zhao, Wei
    Xie, Xiangpeng
    Yue, Dong
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (09): : 5307 - 5316