Leader-Follower Bipartite Output Synchronization on Signed Digraphs Under Adversarial Factors via Data-Based Reinforcement Learning

被引:29
作者
Li, Qin [1 ]
Xia, Lina [1 ]
Song, Ruizhuo [1 ]
Liu, Jian [1 ]
机构
[1] Univ Sci & Technol Beijing, Sch Automat & Elect Engn, Beijing 100083, Peoples R China
基金
中国国家自然科学基金;
关键词
Adversarial inputs; bipartite output synchronization; heterogeneous multiagent systems (MASs); reinforcement learning (RL); resilient H-infinity controller; signed digraphs; MULTIAGENT SYSTEMS; CONTAINMENT CONTROL; FEEDBACK CONTROL; CONSENSUS;
D O I
10.1109/TNNLS.2019.2952611
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The optimal solution to the leader-follower bipartite output synchronization problem is proposed for heterogeneous multiagent systems (MASs) over signed digraphs in the presence of adversarial inputs in this article. For the MASs, the dynamics and dimensions of the followers are different. Distributed observers are first designed to estimate the leader's two-way state and output over signed digraphs. Then, the leader-follower bipartite output synchronization problem on signed graphs is translated into a conventional output distributed leader-follower problem over nonnegative graphs after the state transformation by using the information of followers and observers. The effect of adversarial inputs in sensors or actuators of agents is mitigated by designing the resilient H-infinity controller. A data-based reinforcement learning (RL) algorithm is proposed to obtain the optimal control law, which implies that the dynamics of the followers is not required. Finally, a simulation example is given to verify the effectiveness of the proposed algorithm.
引用
收藏
页码:4185 / 4195
页数:11
相关论文
共 39 条
[21]  
Moghadam R., 2017, 2017 IEEE S SER COMP, DOI [10.1109/SSCI.2017.8280840, DOI 10.1109/SSCI.2017.8280840]
[22]   Resilient Autonomous Control of Distributed Multiagent Systems in Contested Environments [J].
Moghadam, Rohollah ;
Modares, Hamidreza .
IEEE TRANSACTIONS ON CYBERNETICS, 2019, 49 (11) :3957-3967
[23]   Optimal Synchronization Control of Multiagent Systems With Input Saturation via Off-Policy Reinforcement Learning [J].
Qin, Jiahu ;
Li, Man ;
Shi, Yang ;
Ma, Qichao ;
Zheng, Wei Xing .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (01) :85-96
[24]   Robust H∞ Group Consensus for Interacting Clusters of Integrator Agents [J].
Qin, Jiahu ;
Ma, Qichao ;
Zheng, Wei Xing ;
Gao, Huijun ;
Kang, Yu .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2017, 62 (07) :3559-3566
[25]   H∞ Consensus Achievement of Multi-Agent Systems With Directed and Switching Topology Networks [J].
Saboori, Iman ;
Khorasani, Khashayar .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2014, 59 (11) :3104-3109
[26]   Stable value iteration for two-player zero-sum game of discrete-time nonlinear systems based on adaptive dynamic programming [J].
Song, Ruizhuo ;
Zhu, Liao .
NEUROCOMPUTING, 2019, 340 :180-195
[27]   Off-Policy Integral Reinforcement Learning Method to Solve Nonlinear Continuous-Time Multiplayer Nonzero-Sum Games [J].
Song, Ruizhuo ;
Lewis, Frank L. ;
Wei, Qinglai .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (03) :704-713
[28]   Off-Policy Actor-Critic Structure for Optimal Control of Unknown Systems With Disturbances [J].
Song, Ruizhuo ;
Lewis, Frank L. ;
Wei, Qinglai ;
Zhang, Huaguang .
IEEE TRANSACTIONS ON CYBERNETICS, 2016, 46 (05) :1041-1050
[29]  
Vrabie F. L. D. Lewis, 2012, OPTIMAL CONTROL
[30]   Neural Network Based Online Simultaneous Policy Update Algorithm for Solving the HJI Equation in Nonlinear H∞ Control [J].
Wu, Huai-Ning ;
Luo, Biao .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2012, 23 (12) :1884-1895