Soft Actor-Critic-Based Computation Offloading in Multiuser MEC-Enabled IoT-A Lifetime Maximization Perspective

被引:17
作者
Heidarpour, Ali Reza [1 ]
Heidarpour, Mohammad Reza [2 ]
Ardakani, Masoud [1 ]
Tellambura, Chintha [1 ]
Uysal, Murat [3 ]
机构
[1] Univ Alberta, Dept Elect & Comp Engn, Edmonton, AB, Canada
[2] Isfahan Univ Technol, Dept Elect & Comp Engn, Esfahan, Iran
[3] Istanbul Univ, Dept Elect & Elect Engn, TR-34794 Istanbul, Turkiye
基金
加拿大自然科学与工程研究理事会;
关键词
Deep reinforcement learning (DRL); Internet of Things (IoT); lifetime maximization; mobile-edge computing (MEC); soft actor-critic (SAC); RESOURCE-ALLOCATION; POWER ALLOCATION; RELAY SELECTION; EDGE; TIME;
D O I
10.1109/JIOT.2023.3277753
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This article studies the network lifetime optimization problem in a multiuser mobile-edge computing (MEC)-enabled Internet of Things (IoT) system comprising an access point (AP), a MEC server, and a set of K mobile devices (MDs) with limited battery capacity. Considering the residual battery energy at the MDs, stochastic task arrivals, and time-varying wireless fading channels, a soft actor-critic (SAC)-based deep reinforcement learning (DRL) lifetime maximization, called DeepLM, is proposed to jointly optimize the task splitting ratio, the local CPU-cycle frequencies at the MDs, the bandwidth allocation, and the CPU-cycle frequency allocation at the MEC server subject to the task queuing backlogs constraint, the bandwidth constraint, and maximum CPU-cycle frequency constraints at the MDs and the MEC server. Our results reveal that DeepLM enjoys a fast convergence rate and a small oscillation amplitude. We also compare the performance of DeepLM with three benchmark offloading schemes, namely, fully edge computing (FEC), fully local computing (FLC), and random computation offloading (RCO). DeepLM increases the network lifetime by 496% and 229% compared to the FLC and RCO schemes. Interestingly, it achieves such a colossal lifetime improvement when its nonbacklog probability is 0.99, while that of FEC, FLC, and RCO is 0.69, 0.53, and 0.25, respectively, showing a significant performance gain of 30%, 46%, and 74%.
引用
收藏
页码:17571 / 17584
页数:14
相关论文
共 38 条
[1]  
Abramowitz M., 1972, Handbook of Mathematical Functions with Formulas, Graphs, and Mathematical Tables
[2]   Optimized Computation Offloading Performance in Virtual Edge Computing Systems via Deep Reinforcement Learning [J].
Chen, Xianfu ;
Zhang, Honggang ;
Wu, Celimuge ;
Mao, Shiwen ;
Ji, Yusheng ;
Bennis, Mehdi .
IEEE INTERNET OF THINGS JOURNAL, 2019, 6 (03) :4005-4018
[3]   Lifetime Maximization in Mobile Edge Computing Networks [J].
Gupta, Sabyasachi ;
Chakareski, Jacob .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (03) :3310-3321
[4]   Partner Selection Based on Optimal Power Allocation for Lifetime Maximization in Cooperative Networks [J].
Gupta, Sabyasachi ;
Bose, Ranjan .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2017, 66 (04) :3088-3102
[5]  
Haarnoja T, 2018, P MACHINE LEARNING R, V80
[6]  
Haarnoja T, 2019, Arxiv, DOI [arXiv:1812.05905, 10.48550/arXiv.1812.05905]
[7]   Deep Reinforcement Learning for Online Computation Offloading in Wireless Powered Mobile-Edge Computing Networks [J].
Huang, Liang ;
Bi, Suzhi ;
Zhang, Ying-Jun Angela .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2020, 19 (11) :2581-2593
[8]   Relay Selection and Power Allocation for Cooperative Network Based on Energy Pricing [J].
Ke, Feng ;
Feng, Suili ;
Zhuang, Hongcheng .
IEEE COMMUNICATIONS LETTERS, 2010, 14 (05) :396-398
[9]   Deep Reinforcement Learning-Based Adaptive Computation Offloading for MEC in Heterogeneous Vehicular Networks [J].
Ke, Hongchang ;
Wang, Jian ;
Deng, Lingyue ;
Ge, Yuming ;
Wang, Hui .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (07) :7916-7929
[10]  
Kingsbury D, 2015, P1, DOI [10.48550/arXiv.1312.6114, DOI 10.48550/ARXIV.1412.6980, 10.48550/arXiv.1412.6980, DOI 10.1021/bk-2015-1214.ch001]