Prescribed-Time Human-in-the-Loop Optimal Synchronization Control for Multiagent Systems Under DoS Attacks via Reinforcement Learning

被引:0
作者
Huang, Zongsheng [1 ]
Li, Tieshan [1 ,2 ,3 ]
Long, Yue [1 ]
Liang, Hongjing [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Automat Engn, Chengdu 611731, Peoples R China
[2] Univ Elect Sci & Technol China, Yangtze Delta Reg Inst, Huzhou 313000, Peoples R China
[3] Lab Electromagnet Space Cognit & Intelligent Contr, Beijing 100089, Peoples R China
基金
中国国家自然科学基金;
关键词
Topology; Denial-of-service attack; Q-learning; Observers; Synchronization; Switches; Optimal control; Heuristic algorithms; Convergence; Consensus control; Denial-of-service (DoS) attacks; fully distributed control; human-in-the-loop (HiTL) control; neural network (NN)-based reinforcement learning (RL); prescribed-time control (PTC); CONSENSUS; LEADER;
D O I
10.1109/TNNLS.2025.3583248
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The prescribed-time (PT) human-in-the-loop (HiTL) optimal synchronization control problem for multiagent systems (MASs) under link-based denial-of-service (DoS) attacks is investigated. First, the HiTL framework enables the human operator to govern the MASs by transmitting commands to the leader. The link-based DoS attacks cause communication blockages between agents, resulting in topology switching. Under the switching communication topology, a fully distributed observer is proposed for each follower, which simultaneously integrates a prescribed finite-time function to estimate the leader's output within the PT. This observer is characterized by a bounded gain at the PT point and guarantees global practical PT convergence, while avoiding the use of global topology information. By combining the follower dynamics with the proposed observer, an augmented system is developed. Subsequently, the model-free Q-learning algorithm is used to learn the optimal synchronization policy directly from real system data. To reduce computational burden, the Q-learning algorithm is implemented using a single critic neural network (NN) structure, with the least-squares method applied to train the NN weights. The convergence of the Q-functions generated by the proposed Q-learning algorithm is proven. Finally, simulation results verify the effectiveness of the proposed control scheme.
引用
收藏
页数:11
相关论文
共 44 条
[1]   Model-free Q-learning designs for linear discrete-time zero-sum games with application to H-infinity control [J].
Al-Tamimi, Asma ;
Lewis, Frank L. ;
Abu-Khalaf, Murad .
AUTOMATICA, 2007, 43 (03) :473-481
[2]   Human-in-the-Loop Consensus Tracking Control for UAV Systems via an Improved Prescribed Performance Approach [J].
Chen, Lei ;
Liang, Hongjing ;
Pan, Yingnan ;
Li, Tieshan .
IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 2023, 59 (06) :8380-8391
[3]   Novel decentralized adaptive strategies for the synchronization of complex networks [J].
DeLellis, Pietro ;
diBernardo, Mario ;
Garofalo, Francesco .
AUTOMATICA, 2009, 45 (05) :1312-1318
[4]   Data-Driven Tracking Control for Multi-Agent Systems With Unknown Dynamics via Multithreading Iterative Q-Learning [J].
Dong, Tao ;
Gong, Xiaomei ;
Wang, Aijuan ;
Li, Huaqing ;
Huang, Tingwen .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2023, 53 (04) :2533-2542
[5]   Secure Control of Multiagent Systems Against Malicious Attacks: A Brief Survey [J].
He, Wangli ;
Xu, Wenying ;
Ge, Xiaohua ;
Han, Qing-Long ;
Du, Wenli ;
Qian, Feng .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2022, 18 (06) :3595-3608
[6]   Observer-Based Human-in-the-Loop Optimal Output Cluster Synchronization Control for Multiagent Systems: A Model-Free Reinforcement Learning Method [J].
Huang, Zongsheng ;
Li, Tieshan ;
Long, Yue ;
Liang, Hongjing .
IEEE TRANSACTIONS ON CYBERNETICS, 2025, 55 (02) :649-660
[7]   Leader-following consensus of linear time-varying multi-agent systems under fixed and switching topologies [J].
Jiang, Jiahe ;
Jiang, Yangyang .
AUTOMATICA, 2020, 113
[8]  
Kiumarsi B, 2018, IEEE DECIS CONTR P, P3343, DOI 10.1109/CDC.2018.8618994
[9]  
Li RH, 2025, IEEE T SYST MAN CY-S, V55, P3188, DOI [10.1109/tsmc.2025.3539656, 10.1109/TSMC.2025.3539656]
[10]   Designing Fully Distributed Consensus Protocols for Linear Multi-Agent Systems With Directed Graphs [J].
Li, Zhongkui ;
Wen, Guanghui ;
Duan, Zhisheng ;
Ren, Wei .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2015, 60 (04) :1152-1157