Dependent Task Offloading in Edge Computing Using GNN and Deep Reinforcement Learning

被引:12
作者
Cao, Zequn [1 ]
Deng, Xiaoheng [1 ]
Yue, Sheng [2 ]
Jiang, Ping [1 ]
Ren, Ju [2 ]
Gui, Jinsong [1 ]
机构
[1] Cent South Univ, Sch Comp Sci & Engn, Changsha 410075, Peoples R China
[2] Tsinghua Univ, Dept Comp Sci & Technol, BNRist, Beijing 100084, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Training; Heuristic algorithms; Cloud computing; Quality of experience; Scheduling; Program processors; Deep reinforcement learning (DRL); directed acyclic graph (DAG); edge computing (EC); graph attention network (GAT); task offloading; RESOURCE-ALLOCATION;
D O I
10.1109/JIOT.2024.3374969
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Task offloading is a widely used technology in edge computing (EC), which declines the makespan of user task with the aid of resourceful edge servers. How to solve the competition for computation and communication resources among tasks is a fundamental issue in task offloading. Besides, real-life user tasks often comprise multiple interdependent subtasks. Dependencies among subtasks significantly raises the complexity of task offloading, and makes it difficult to propose generalized approaches for scenarios of different sizes. In this article, we study the dependent task offloading (DTO) problem within both single-user single-edge and multiuser multiedge scenario. First, we use directed acyclic graph (DAG) to model dependent task, where nodes and directed edges represent the subtasks and their interdependencies, respectively. Then, we propose a task scheduling method based on graph attention network (GAT) and deep reinforcement learning (DRL) to minimize the makespan of user tasks. More specifically, our method introduces a multidiscrete action DRL scheduler that simultaneously determines which subtask to consider and whether it should be offloaded at each step, and employs GAT to encode the graph-based state representation. To stabilize and speed up DRL scheduler training, we pretrain GAT encoder with unsupervised learning. Extensive experiments demonstrate that our proposed approach can be applied to various environments and outperforms prior methods.
引用
收藏
页码:21632 / 21646
页数:15
相关论文
共 49 条
[11]  
Hessel M., 2018, PROC AAAI C ARTIF, P1
[12]   Multi-Hop Cooperative Computation Offloading for Industrial IoT-Edge-Cloud Computing Environments [J].
Hong, Zicong ;
Chen, Wuhui ;
Huang, Huawei ;
Guo, Song ;
Zheng, Zibin .
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2019, 30 (12) :2759-2774
[13]  
Huang S., 2022, The 37 Implementation Details of Proximal Policy Optimization
[14]  
Lan C.L., 2022, arXiv
[15]  
Lillicrap T. P., 2019, Continuous control with deep reinforcement learning
[16]   Towards Unsupervised Deep Graph Structure Learning [J].
Liu, Yixin ;
Zheng, Yu ;
Zhang, Daokun ;
Chen, Hongxu ;
Peng, Hao ;
Pan, Shirui .
PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, :1392-1403
[17]   Resource Scheduling in Edge Computing: A Survey [J].
Luo, Quyuan ;
Hu, Shihong ;
Li, Changle ;
Li, Guanghui ;
Shi, Weisong .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2021, 23 (04) :2131-2165
[18]   Mobile Edge Computing: A Survey on Architecture and Computation Offloading [J].
Mach, Pavel ;
Becvar, Zdenek .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2017, 19 (03) :1628-1656
[19]  
Madjiheurem S., 2019, PROC 22 INT C A, P3391
[20]   Learning Scheduling Algorithms for Data Processing Clusters [J].
Mao, Hongzi ;
Schwarzkopf, Malte ;
Venkatakrishnan, Shaileshh Bojja ;
Meng, Zili ;
Alizadeh, Mohammad .
SIGCOMM '19 - PROCEEDINGS OF THE ACM SPECIAL INTEREST GROUP ON DATA COMMUNICATION, 2019, :270-288