Optimal Containment Control of Unknown Heterogeneous Systems With Active Leaders

被引:89
作者
Yang, Yongliang [1 ,2 ]
Modares, Hamidreza [3 ]
Wunsch, Donald C., II [3 ]
Yin, Yixin [1 ,2 ]
机构
[1] Univ Sci & Technol Beijing, Sch Automat & Elect Engn, Beijing 100083, Peoples R China
[2] Univ Sci & Technol Beijing, Key Lab Knowledge Automat Ind Proc, Minist Educ, Beijing 100083, Peoples R China
[3] Missouri Univ Sci & Technol, Dept Elect & Comp Engn, Rolla, MO 65409 USA
基金
中国国家自然科学基金; 美国国家科学基金会;
关键词
Active leader; containment control; distributed observer; heterogeneous systems; model free; reinforcement learning (RL); LINEAR MULTIAGENT SYSTEMS; CONTINUOUS-TIME SYSTEMS; TRACKING CONTROL; OUTPUT SYNCHRONIZATION; NETWORK; CONSENSUS;
D O I
10.1109/TCST.2018.2794336
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This brief presents a partially model-free solution to the distributed containment control of multiagent systems using off-policy reinforcement learning (RL). The followers are assumed to be heterogeneous with different dynamics, and the leaders are assumed to be active in the sense that their control inputs can be nonzero. Optimality is explicitly imposed in solving the containment problem to not only drive the agents' states into a convex hull of the leaders' states but also minimize their transient responses. Inhomogeneous algebraic Riccati equations (AREs) are derived to solve the optimal containment control with active leaders. The resulting control protocol for each agent depends on its own state and an estimation of an interior point inside the convex hull spanned by the leaders. This estimation is provided by designing a distributed observer for a trajectory inside the convex hull of active leaders. Only the knowledge of the leaders' dynamics is required by the observer. An off-policy RL algorithm is developed to solve the inhomogeneous AREs online in real time without requiring any knowledge of the followers' dynamics. Finally, a simulation example is presented to show the effectiveness of the presented algorithm.
引用
收藏
页码:1228 / 1236
页数:9
相关论文
共 36 条
[1]   Distributed Containment Control for Multiple Autonomous Vehicles With Double-Integrator Dynamics: Algorithms and Experiments [J].
Cao, Yongcan ;
Stuart, Daniel ;
Ren, Wei ;
Meng, Ziyang .
IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2011, 19 (04) :929-938
[2]   On the Internal Model Principle in the Coordination of Nonlinear Systems [J].
De Persis, Claudio ;
Jayawardhana, Bayu .
IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2014, 1 (03) :272-282
[3]  
Engwerda J., 2005, LQ dynamic optimization and differential games
[4]   Containment control of heterogeneous linear multi-agent systems [J].
Haghshenas, Hamed ;
Badamchizadeh, Mohammad Ali ;
Baradarannia, Mandi .
AUTOMATICA, 2015, 54 :210-216
[5]   Adaptive Neural Network Control of a Flapping Wing Micro Aerial Vehicle With Disturbance Observer [J].
He, Wei ;
Yan, Zichen ;
Sun, Changyin ;
Chen, Yunan .
IEEE TRANSACTIONS ON CYBERNETICS, 2017, 47 (10) :3452-3465
[6]   Control Design for Nonlinear Flexible Wings of a Robotic Aircraft [J].
He, Wei ;
Zhang, Shuang .
IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2017, 25 (01) :351-357
[7]   Multiagent Information Fusion and Cooperative Control in Target Search [J].
Hu, Jinwen ;
Xie, Lihua ;
Lum, Kai-Yew ;
Xu, Jun .
IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2013, 21 (04) :1223-1235
[8]   Robust Output Synchronization of a Network of Heterogeneous Nonlinear Agents Via Nonlinear Regulation Theory [J].
Isidori, Alberto ;
Marconi, Lorenzo ;
Casadei, Giacomo .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2014, 59 (10) :2680-2691
[9]   Computational adaptive optimal control for continuous-time linear systems with completely unknown dynamics [J].
Jiang, Yu ;
Jiang, Zhong-Ping .
AUTOMATICA, 2012, 48 (10) :2699-2704
[10]   Model-Based Reinforcement Learning in Differential Graphical Games [J].
Kamalapurkar, Rushikesh ;
Klotz, Justin R. ;
Walters, Patrick ;
Dixon, Warren E. .
IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2018, 5 (01) :423-433