Deep reinforcement learning for the computation offloading in MIMO-based Edge Computing

被引:24
作者
Sadiki, Abdeladim [1 ]
Bentahar, Jamal [1 ,3 ]
Dssouli, Rachida [1 ]
En-Nouaary, Abdeslam [2 ]
Otrok, Hadi [3 ]
机构
[1] Concordia Univ, Concordia Inst Informat Syst Engn, Montreal, PQ, Canada
[2] Inst Natl Postes Telecommun, STRS Lab, Rabat, Morocco
[3] Khalifa Univ, Dept EECS, Abu Dhabi, U Arab Emirates
基金
加拿大自然科学与工程研究理事会;
关键词
Multi-access Edge Computing; Massive multiple-input multiple-output; Deep reinforcement learning; Computation offloading; RESOURCE-ALLOCATION; MULTIPLE-ACCESS; MOBILE; TASK; IMPACT;
D O I
10.1016/j.adhoc.2022.103080
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-access Edge Computing (MEC) has recently emerged as a potential technology to serve the needs of mobile devices (MDs) in 5G and 6G cellular networks. By offloading tasks to high-performance servers installed at the edge of the wireless networks, resource-limited MDs can cope with the proliferation of the recent computationally-intensive applications. In this paper, we study the computation offloading problem in a massive multiple-input multiple-output (MIMO)-based MEC system where the base stations are equipped with a large number of antennas. Our objective is to minimize the power consumption and offloading delay at the MDs under the stochastic system environment. To this end, we introduce new formulation of the problem as a Markov Decision Process (MDP) and propose two Deep Reinforcement Learning (DRL) algorithms to learn the optimal offloading policy without any prior knowledge of the environment dynamics. First, a Deep Q-Network (DQN)-based algorithm to solve the curse of the state space explosion is defined. Then, a more general Proximal Policy Optimization (PPO)-based algorithm to solve the problem of discrete action space is introduced. Simulation results show that our DRL-based solutions outperform the state-of-the-art algorithms. Moreover, our PPO algorithm exhibits stable performance and efficient offloading results compared to the benchmarks DQN and Double DQN (DDQN) strategies.
引用
收藏
页数:15
相关论文
共 62 条
[1]   Mobile Edge Computing: A Survey [J].
Abbas, Nasir ;
Zhang, Yan ;
Taherkordi, Amir ;
Skeie, Tor .
IEEE INTERNET OF THINGS JOURNAL, 2018, 5 (01) :450-465
[2]   Target localization using Multi-Agent Deep Reinforcement Learning with Proximal Policy Optimization [J].
Alagha, Ahmed ;
Singh, Shakti ;
Mizouni, Rabeb ;
Bentahar, Jamal ;
Otrok, Hadi .
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2022, 136 :342-357
[3]   An overview of massive MIMO localization techniques in wireless cellular networks: Recent advances and outlook [J].
Alamu, Olumide ;
Iyaomolere, Babatunde ;
Abdulrahman, Abdulfatai .
AD HOC NETWORKS, 2021, 111
[4]  
Ale L., 2021, IEEE T COGN COMMUN
[5]   Task Offloading and Resource Allocation for Mobile Edge Computing by Deep Reinforcement Learning Based on SARSA [J].
Alfakih, Taha ;
Hassan, Mohammad Mehedi ;
Gumaei, Abdu ;
Savaglio, Claudio ;
Fortino, Giancarlo .
IEEE ACCESS, 2020, 8 :54074-54084
[6]  
[Anonymous], 2020, MASSIVE MIMO CHANNEL
[7]   Delay Characterization of Mobile-Edge Computing for 6G Time-Sensitive Services [J].
Cao, Jianyu ;
Feng, Wei ;
Ge, Ning ;
Lu, Jianhua .
IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (05) :3758-3773
[8]   Massive MIMO Systems for 5G and beyond Networks-Overview, Recent Trends, Challenges, and Future Research Direction [J].
Chataut, Robin ;
Akl, Robert .
SENSORS, 2020, 20 (10)
[9]   Task Offloading for Mobile Edge Computing in Software Defined Ultra-Dense Network [J].
Chen, Min ;
Hao, Yixue .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2018, 36 (03) :587-597
[10]  
Chen X., 2018, 88 IEEE VEHICULAR TE, P1