Reinforcement Learning for Synchronization of Heterogeneous Multiagent Systems by Improved Q-Functions

被引:0
|
作者
Li, Jinna [1 ]
Yuan, Lin [1 ]
Cheng, Weiran [1 ]
Chai, Tianyou [2 ]
Lewis, Frank L. [3 ]
机构
[1] Liaoning Petrochem Univ, Sch Informat & Control Engn, Fushun 113001, Liaoning, Peoples R China
[2] Northeastern Univ, State Key Lab Synthet Automat Proc Ind, Shenyang 110819, Peoples R China
[3] Univ Texas Arlington, UTA Res Inst, Arlington, TX 76118 USA
基金
中国国家自然科学基金;
关键词
Synchronization; Protocols; Heuristic algorithms; Decision making; Nash equilibrium; Multi-agent systems; Games; Data-driven control; distributed control; multiagent systems (MASs); reinforcement learning (RL); synchronization;
D O I
10.1109/TCYB.2024.3440333
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This article dedicates to investigating a methodology for enhancing adaptability to environmental changes of reinforcement learning (RL) techniques with data efficiency, by which a joint control protocol is learned using only data for multiagent systems (MASs). Thus, all followers are able to synchronize themselves with the leader and minimize their individual performance. To this end, an optimal synchronization problem of heterogeneous MASs is first formulated, and then an arbitration RL mechanism is developed for well addressing key challenges faced by the current RL techniques, that is, insufficient data and environmental changes. In the developed mechanism, an improved Q-function with an arbitration factor is designed for accommodating the fact that control protocols tend to be made by historic experiences and instinctive decision-making, such that the degree of control over agents' behaviors can be adaptively allocated by on-policy and off-policy RL techniques for the optimal multiagent synchronization problem. Finally, an arbitration RL algorithm with critic-only neural networks is proposed, and theoretical analysis and proofs of synchronization and performance optimality are provided. Simulation results verify the effectiveness of the proposed method.
引用
收藏
页码:6545 / 6558
页数:14
相关论文
共 50 条
  • [31] Decentralized Reinforcement Learning Inspired by Multiagent Systems
    Adjodah, Dhaval
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), 2018, : 1729 - 1730
  • [32] Scheduling in Multiagent Systems Using Reinforcement Learning
    I. K. Minashina
    R. A. Gorbachev
    E. M. Zakharova
    Doklady Mathematics, 2022, 106 : S70 - S78
  • [33] Data-Based Optimal Synchronization Control for Discrete-Time Nonlinear Heterogeneous Multiagent Systems
    Fu, Hao
    Chen, Xin
    Wang, Wei
    Wu, Min
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (04) : 2477 - 2490
  • [34] Model-Free Event-Triggered Consensus Algorithm for Multiagent Systems Using Reinforcement Learning Method
    Long, Mingkang
    Su, Housheng
    Zeng, Zhigang
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (08): : 5212 - 5221
  • [35] Data-Based Output Synchronization of Discrete-Time Heterogeneous Multiagent Systems With Sensor Faults
    Liu, Yingying
    Wang, Zhanshan
    IEEE TRANSACTIONS ON CYBERNETICS, 2024, 54 (01) : 265 - 272
  • [36] Leader-Follower Output Synchronization of Linear Heterogeneous Systems With Active Leader Using Reinforcement Learning
    Yang, Yongliang
    Modares, Hamidreza
    Wunsch, Donald C., II
    Yin, Yixin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (06) : 2139 - 2153
  • [37] A Low Complexity Robust Output Synchronization Protocol With Prescribed Performance for High-Order Heterogeneous Uncertain MIMO Nonlinear Multiagent Systems
    Katsoukis, Ilias
    Rovithakis, George A.
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (06) : 3128 - 3133
  • [38] Hierarchical Optimal Synchronization for Linear Systems via Reinforcement Learning: A Stackelberg-Nash Game Perspective
    Li, Man
    Qin, Jiahu
    Ma, Qichao
    Zheng, Wei Xing
    Kang, Yu
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (04) : 1600 - 1611
  • [39] Distributed learning consensus control based on neural networks for heterogeneous nonlinear multiagent systems
    Shen, Dong
    Zhang, Chao
    Xu, Jian-Xin
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2019, 29 (13) : 4328 - 4347
  • [40] Inverse Reinforcement Learning for Decentralized Non-Cooperative Multiagent Systems
    Reddy, Tummalapalli Sudhamsh
    Gopikrishna, Vamsikrishna
    Zaruba, Gergely
    Huber, Manfred
    PROCEEDINGS 2012 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2012, : 1930 - 1935