Behavioral control task supervisor with memory based on reinforcement learning for human-multi-robot coordination systems

被引:6
作者
Huang, Jie [1 ,2 ,3 ]
Mo, Zhibin [1 ,2 ,3 ]
Zhang, Zhenyi [1 ,2 ,3 ]
Chen, Yutao [1 ,2 ,3 ]
机构
[1] Fuzhou Univ, Sch Elect Engn & Automat, Fuzhou 350108, Peoples R China
[2] Fuzhou Univ, 5G Ind Internet Inst, Fuzhou 350108, Peoples R China
[3] Fuzhou Univ, Key Lab Ind Automat Control Technol & Informat Pr, Fuzhou 350108, Peoples R China
基金
中国国家自然科学基金;
关键词
Human-multi-robot coordination systems; Null-space-based behavioral control; Task supervisor; Reinforcement learning; Knowledge base; TP18; DECISION-MAKING; COLLABORATION; AGENTS;
D O I
10.1631/FITEE.2100280
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this study, a novel reinforcement learning task supervisor (RLTS) with memory in a behavioral control framework is proposed for human-multi-robot coordination systems (HMRCSs). Existing HMRCSs suffer from high decision-making time cost and large task tracking errors caused by repeated human intervention, which restricts the autonomy of multi-robot systems (MRSs). Moreover, existing task supervisors in the null-space-based behavioral control (NSBC) framework need to formulate many priority-switching rules manually, which makes it difficult to realize an optimal behavioral priority adjustment strategy in the case of multiple robots and multiple tasks. The proposed RLTS with memory provides a detailed integration of the deep Q-network (DQN) and long short-term memory (LSTM) knowledge base within the NSBC framework, to achieve an optimal behavioral priority adjustment strategy in the presence of task conflict and to reduce the frequency of human intervention. Specifically, the proposed RLTS with memory begins by memorizing human intervention history when the robot systems are not confident in emergencies, and then reloads the history information when encountering the same situation that has been tackled by humans previously. Simulation results demonstrate the effectiveness of the proposed RLTS. Finally, an experiment using a group of mobile robots subject to external noise and disturbances validates the effectiveness of the proposed RLTS with memory in uncertain real-world environments.
引用
收藏
页码:1174 / 1188
页数:15
相关论文
共 50 条
[41]   Trajectory Control of An Articulated Robot Based on Direct Reinforcement Learning [J].
Tsai, Chia-Hao ;
Lin, Jun-Ji ;
Hsieh, Teng-Feng ;
Yen, Jia-Yush .
ROBOTICS, 2022, 11 (05)
[42]   Adaptive Admittance Control for Physical Human-Robot Interaction based on Imitation and Reinforcement Learning [J].
Guo, Mou ;
Yao, Bitao ;
Ji, Zhenrui ;
Xu, Wenjun ;
Zhou, Zude .
2023 29TH INTERNATIONAL CONFERENCE ON MECHATRONICS AND MACHINE VISION IN PRACTICE, M2VIP 2023, 2023,
[43]   Evolving hierarchical memory-prediction machines in multi-task reinforcement learning [J].
Kelly, Stephen ;
Voegerl, Tatiana ;
Banzhaf, Wolfgang ;
Gondro, Cedric .
GENETIC PROGRAMMING AND EVOLVABLE MACHINES, 2021, 22 (04) :573-605
[44]   Evolving hierarchical memory-prediction machines in multi-task reinforcement learning [J].
Stephen Kelly ;
Tatiana Voegerl ;
Wolfgang Banzhaf ;
Cedric Gondro .
Genetic Programming and Evolvable Machines, 2021, 22 :573-605
[45]   Optimal formation tracking control based on reinforcement learning for multi-UAV systems [J].
Wang, Weizhen ;
Chen, Xin ;
Jia, Jiangbo ;
Wu, Kaili ;
Xie, Mingyang .
CONTROL ENGINEERING PRACTICE, 2023, 141
[46]   Multi-Agent Reinforcement Learning Based Actuator Control for EV HVAC Systems [J].
Joo, Sungho ;
Lee, Dongmin ;
Kim, Minseop ;
Lee, Taeho ;
Choi, Sanghyeok ;
Kim, Seungju ;
Lee, Jeyeol ;
Kim, Joongjae ;
Lim, Yongsub ;
Lee, Jeonghoon .
IEEE ACCESS, 2023, 11 :7574-7587
[47]   A multi-task deep reinforcement learning framework based on curriculum learning and policy distillation for quadruped robot motor skill training [J].
Chen, Liang ;
Shen, Bo ;
Hong, Jiale .
SYSTEMS SCIENCE & CONTROL ENGINEERING, 2025, 13 (01)
[48]   Multi-Agent Reinforcement Learning based on K-Means Clustering in Multi-Robot Cooperative Systems [J].
Liu Chang-an ;
Liu Fei ;
Liu Chun-yang ;
Wu Hua .
OPTICAL, ELECTRONIC MATERIALS AND APPLICATIONS, PTS 1-2, 2011, 216 :75-80
[49]   Multi-Task Reinforcement Learning Based on Parallel Recombination Networks [J].
Liu, Manlu ;
Zhang, Qingbo ;
Qian, Weimin .
IEEE ACCESS, 2025, 13 :80113-80122
[50]   Curriculum-Based Asymmetric Multi-Task Reinforcement Learning [J].
Huang, Hanchi ;
Ye, Deheng ;
Shen, Li ;
Liu, Wei .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (06) :7258-7269