Distributed deep reinforcement learning for independent task offloading in Mobile Edge Computing

被引:0
作者
Darchini-Tabrizi, Mohsen [1 ]
Roudgar, Amirhossein [1 ]
Entezari-Maleki, Reza [1 ,2 ,3 ]
Sousa, Leonel [3 ]
机构
[1] Iran Univ Sci & Technol, Sch Comp Engn, Tehran, Iran
[2] Inst Res Fundamental Sci IPM, Sch Comp Sci, Tehran, Iran
[3] Univ Lisbon, INESC ID, Inst Super Tecn, Lisbon, Portugal
关键词
Mobile Edge Computing; Task offloading; Task size prediction; Deep reinforcement learning;
D O I
10.1016/j.jnca.2025.104211
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Mobile Edge Computing (MEC) has been identified as an innovative paradigm to improve the performance and efficiency of mobile applications by offloading computation-intensive tasks to nearby edge servers. However, the effective implementation of task offloading in MEC systems faces challenges due to uncertainty, heterogeneity, and dynamicity. Deep Reinforcement Learning (DRL) provides a powerful approach for devising optimal task offloading policies in complex and uncertain environments. This paper presents a DRL-based task offloading approach using Deep Deterministic Policy Gradient (DDPG) and Distributed Distributional Deep Deterministic Policy Gradient (D4PG) algorithms. The proposed solution establishes a distributed system, where multiple mobile devices act as Reinforcement Learning (RL) agents to optimize their individual performance. To reduce the computational complexity of the neural networks, Gated Recurrent Units (GRU) are used instead of Long Short-Term Memory (LSTM) units to predict the load of edge nodes within the observed state. In addition, a GRU-based sequencing model is introduced to estimate task sizes in specific scenarios where these sizes are unknown. Finally, a novel scheduling algorithm is proposed that outperforms commonly used approaches by leveraging the estimated task sizes to achieve superior performance. Comprehensive simulations were conducted to evaluate the efficacy of the proposed approach, benchmarking it against multiple baseline and state-of-the-art algorithms. Results show significant improvements in terms of average processing delay and task drop rates, thereby confirming the success of the proposed approach.
引用
收藏
页数:16
相关论文
共 56 条
[1]   Meta-heuristic-based offloading task optimization in mobile edge computing [J].
Abbas, Aamir ;
Raza, Ali ;
Aadil, Farhan ;
Maqsood, Muazzam .
INTERNATIONAL JOURNAL OF DISTRIBUTED SENSOR NETWORKS, 2021, 17 (06)
[2]  
Abu-Taleb N.A., 2022, INT C EM SMART TECHN, P1
[3]   Efficient Task Offloading Strategy for Energy-Constrained Edge Computing Environments: A Hybrid Optimization Approach [J].
Alsadie, Deafallah .
IEEE ACCESS, 2024, 12 :85089-85102
[4]   Lyapunov-Guided Deep Reinforcement Learning for Stable Online Computation Offloading in Mobile-Edge Computing Networks [J].
Bi, Suzhi ;
Huang, Liang ;
Wang, Hui ;
Zhang, Ying-Jun Angela .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (11) :7519-7537
[5]  
Carmo P.R.X., 2024, J. Netw. Comput. Appl., V232
[6]   DRL based partial offloading for maximizing sum computation rate of FDMA-based wireless powered mobile edge computing [J].
Chen, Wenchao ;
Shen, Guanqun ;
Chi, Kaikai ;
Zhang, Shubin ;
Chen, Xiaolong .
COMPUTER NETWORKS, 2022, 214
[7]   Task Offloading and Resource Pricing Based on Game Theory in UAV-Assisted Edge Computing [J].
Chen, Zhuoyue ;
Yang, Yaozong ;
Xu, Jiajie ;
Chen, Ying ;
Huang, Jiwei .
IEEE TRANSACTIONS ON SERVICES COMPUTING, 2025, 18 (01) :440-452
[8]   Computation offloading in mobile edge computing networks: A survey [J].
Feng, Chuan ;
Han, Pengchao ;
Zhang, Xu ;
Yang, Bowen ;
Liu, Yejun ;
Guo, Lei .
JOURNAL OF NETWORK AND COMPUTER APPLICATIONS, 2022, 202
[9]   Deep Reinforcement Learning and Markov Decision Problem for Task Offloading in Mobile Edge Computing [J].
Gao, Xiaohu ;
Ang, Mei Choo ;
Althubiti, Sara A. .
JOURNAL OF GRID COMPUTING, 2023, 21 (04)
[10]   A survey on computation offloading and service placement in fog computing-based IoT [J].
Gasmi, Kaouther ;
Dilek, Selma ;
Tosun, Suleyman ;
Ozdemir, Suat .
JOURNAL OF SUPERCOMPUTING, 2022, 78 (02) :1983-2014