Hierarchical Deep Reinforcement Learning for Computation Offloading in Autonomous Multi-Robot Systems

被引:1
作者
Gao, Wen [1 ]
Yu, Zhiwen [1 ]
Wang, Liang [1 ]
Cui, Helei [1 ]
Guo, Bin [1 ]
Xiong, Hui [2 ]
机构
[1] Northwestern Polytech Univ, Sch Comp Sci, Xian 710072, Peoples R China
[2] Hong Kong Univ Sci & Technol Guangzhou, Thust Artificial Intelligence, Guangzhou 511453, Peoples R China
基金
中国国家自然科学基金;
关键词
Robots; Graphics processing units; Resource management; Computational modeling; Loading; Processor scheduling; Load modeling; Delays; Deep reinforcement learning; Collaboration; Computation offloading; multi-robot systems; reinforcement learning;
D O I
10.1109/LRA.2024.3511408
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
To ensure system responsiveness, some compute-intensive tasks are usually offloaded to cloud or edge computing devices. In environments where connection to external computing facilities is unavailable, computation offloading among members within an autonomous multi-robot system (AMRS) becomes a solution. The challenge lies in how to maximize the use of other members' idle resources without disrupting their local computation tasks. Therefore, this study proposes HRL-AMRS, a hierarchical deep reinforcement learning framework designed to distribute computational loads and reduce the processing time of computational tasks within an AMRS. In this framework, the high-level must consider the impact of data loading scales determined by low-level under varying computational device states on the actual processing times. In addition, the low-level employs Long Short-Term Memory (LSTM) networks to enhance the understanding of time-series states of computing devices. Experimental results show that, across various task sizes and numbers of robots, the framework reduces processing times by an average of 4.32% compared to baseline methods.
引用
收藏
页码:540 / 547
页数:8
相关论文
共 21 条
[1]   Dynamic Task Allocation for Robotic Edge System Resilience Using Deep Reinforcement Learning [J].
Afrin, Mahbuba ;
Jin, Jiong ;
Rahman, Ashfaqur ;
Li, Shi ;
Tian, Yu-Chu ;
Li, Yan .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (03) :1438-1450
[2]   Developing Real-Time Scheduling Policy by Deep Reinforcement Learning [J].
Bo, Zitong ;
Qiao, Ying ;
Leng, Chang ;
Wang, Hongan ;
Guo, Chaoping ;
Zhang, Shaohui .
2021 IEEE 27TH REAL-TIME AND EMBEDDED TECHNOLOGY AND APPLICATIONS SYMPOSIUM (RTAS 2021), 2021, :131-142
[3]   Network offloading policies for cloud robotics: a learning-based approach [J].
Chinchali, Sandeep ;
Sharma, Apoorva ;
Harrison, James ;
Elhafsi, Amine ;
Kang, Daniel ;
Pergament, Evgenya ;
Cidon, Eyal ;
Katti, Sachin ;
Pavone, Marco .
AUTONOMOUS ROBOTS, 2021, 45 (07) :997-1012
[4]   Multi-Agent Deep Reinforcement Learning-Based Interdependent Computing for Mobile Edge Computing-Assisted Robot Teams [J].
Cui, Qimei ;
Zhao, Xiyu ;
Ni, Wei ;
Hu, Zheng ;
Tao, Xiaofeng ;
Zhang, Ping .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (05) :6599-6610
[5]   UAV-Assisted Task Offloading in Vehicular Edge Computing Networks [J].
Dai, Xingxia ;
Xiao, Zhu ;
Jiang, Hongbo ;
Lui, John C. S. .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (04) :2520-2534
[6]  
Foerster JN, 2018, AAAI CONF ARTIF INTE, P2974
[7]   Kalmia: A Heterogeneous QoS-aware Scheduling Framework for DNN Tasks on Edge Servers [J].
Fu, Ziyan ;
Ren, Ju ;
Zhang, Deyu ;
Zhou, Yuezhi ;
Zhang, Yaoxue .
IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2022), 2022, :780-789
[8]   InSS: An Intelligent Scheduling Orchestrator for Multi-GPU Inference With Spatio-Temporal Sharing [J].
Han, Ziyi ;
Zhou, Ruiting ;
Xu, Chengzhong ;
Zeng, Yifan ;
Zhang, Renli .
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2024, 35 (10) :1735-1748
[9]   Edge Computing in 5G for Drone Navigation: What to Offload? [J].
Hayat, Samira ;
Jung, Roland ;
Hellwagner, Hermann ;
Bettstetter, Christian ;
Emini, Driton ;
Schnieders, Dominik .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) :2571-2578
[10]   Voronoi-Based Multi-Robot Autonomous Exploration in Unknown Environments via Deep Reinforcement Learning [J].
Hu, Junyan ;
Niu, Hanlin ;
Carrasco, Joaquin ;
Lennox, Barry ;
Arvin, Farshad .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (12) :14413-14423