Prescribed-time formation tracking in multi-agent systems via reinforcement learning-based hybrid impulsive control with time delays

被引:1
作者
Liang, Zhanlue
Gu, Yanlin [1 ]
Li, Ping [2 ]
Tao, Yiwen [3 ]
机构
[1] Northwestern Polytech Univ, Sch Elect & Informat, Xian, Peoples R China
[2] Southwest Minzu Univ, Sch Math, Chengdu 610041, Sichuan, Peoples R China
[3] Zhengzhou Univ, Sch Math & Stat, Zhengzhou 450001, Henan, Peoples R China
关键词
Prescribed-time formation tracking; Hybrid impulsive control; Multi-agent system; Reinforcement learning; Time delays; Average impulsive interval; TO-STATE STABILITY; FLOCKING CONTROL; STABILIZATION; CONSENSUS;
D O I
10.1016/j.eswa.2025.126723
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper addresses the prescribed-time formation stabilization in nonlinear multi-agent systems using a novel reinforcement learning-based hybrid impulsive control framework that incorporates delayed control impulses. The approach leverages Lyapunov functionals, impulsive comparison theory, average impulsive interval methods, and graph theory to derive sufficient conditions for achieving prescribed-time formation stabilization. These conditions are formulated in terms of continuous and impulsive feedback gains, time delay durations, and average impulsive interval lengths. Importantly, the inclusion of stabilizing control impulses counteracts the destabilizing effects of continuous dynamics. Additionally, deep reinforcement learning techniques are employed to optimize the impulsive control sequence, aiming to maximize rewards derived from the control objectives and system states. Numerical simulation examples are presented to demonstrate the effectiveness and validity of the proposed analytical results, providing comparative assessments of overall control performance.
引用
收藏
页数:16
相关论文
共 59 条
[31]   A unified synchronization criterion for impulsive dynamical networks [J].
Lu, Jianquan ;
Ho, Daniel W. C. ;
Cao, Jinde .
AUTOMATICA, 2010, 46 (07) :1215-1221
[32]  
Mnih V, 2013, Arxiv, DOI [arXiv:1312.5602, 10.48550/arXiv.1312.5602, DOI 10.48550/ARXIV.1312.5602]
[33]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533
[34]   Variable resolution discretization in optimal control [J].
Munos, R ;
Moore, A .
MACHINE LEARNING, 2002, 49 (2-3) :291-323
[35]   Fixed-Time and Prescribed-Time Consensus Control of Multiagent Systems and Its Applications: A Survey of Recent Trends and Methodologies [J].
Ning, Boda ;
Han, Qing-Long ;
Zuo, Zongyu ;
Ding, Lei ;
Lu, Qiang ;
Ge, Xiaohua .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2023, 19 (02) :1121-1135
[36]   Multi-agent robot systems as distributed autonomous systems [J].
Ota, J .
ADVANCED ENGINEERING INFORMATICS, 2006, 20 (01) :59-70
[37]   Data-driven distributed formation control of under-actuated unmanned surface vehicles with collision avoidance via model-based deep reinforcement learning [J].
Pan, Chao ;
Peng, Zhouhua ;
Liu, Lu ;
Wang, Dan .
OCEAN ENGINEERING, 2023, 267
[38]   Hysteresis quantified control for switched reaction-diffusion systems and its application [J].
Peng, Zenglong ;
Song, Xiaona ;
Song, Shuai ;
Stojanovic, Vladimir .
COMPLEX & INTELLIGENT SYSTEMS, 2023, 9 (06) :7451-7460
[39]   Reward is enough [J].
Silver, David ;
Singh, Satinder ;
Precup, Doina ;
Sutton, Richard S. .
ARTIFICIAL INTELLIGENCE, 2021, 299
[40]  
Sutton RS, 2018, ADAPT COMPUT MACH LE, P1