Deep Reinforcement Learning for Task Offloading in Mobile Edge Computing Systems

被引:324
作者
Tang, Ming [1 ]
Wong, Vincent W. S. [1 ]
机构
[1] Univ British Columbia, Dept Elect & Comp Engn, Vancouver, BC V6T 1Z4, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Task analysis; Mobile handsets; Delays; Heuristic algorithms; Mobile computing; Edge computing; Distributed algorithms; Mobile edge computing; computation offloading; resource allocation; deep reinforcement learning; deep Q-learning; RESOURCE-ALLOCATION; NETWORKS;
D O I
10.1109/TMC.2020.3036871
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In mobile edge computing systems, an edge node may have a high load when a large number of mobile devices offload their tasks to it. Those offloaded tasks may experience large processing delay or even be dropped when their deadlines expire. Due to the uncertain load dynamics at the edge nodes, it is challenging for each device to determine its offloading decision (i.e., whether to offload or not, and which edge node it should offload its task to) in a decentralized manner. In this work, we consider non-divisible and delay-sensitive tasks as well as edge load dynamics, and formulate a task offloading problem to minimize the expected long-term cost. We propose a model-free deep reinforcement learning-based distributed algorithm, where each device can determine its offloading decision without knowing the task models and offloading decision of other devices. To improve the estimation of the long-term cost in the algorithm, we incorporate the long short-term memory (LSTM), dueling deep Q-network (DQN), and double-DQN techniques. Simulation results show that our proposed algorithm can better exploit the processing capacities of the edge nodes and significantly reduce the ratio of dropped tasks and average delay when compared with several existing algorithms.
引用
收藏
页码:1985 / 1997
页数:13
相关论文
共 34 条
[21]   Stochastic Joint Radio and Computational Resource Management for Multi-User Mobile-Edge Computing Systems [J].
Mao, Yuyi ;
Zhang, Jun ;
Song, S. H. ;
Letaief, Khaled B. .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2017, 16 (09) :5994-6009
[22]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533
[23]   ULOOF: A User Level Online Offloading Framework for Mobile Edge Computing [J].
Neto, Jose Leal D. ;
Yu, Se-Young ;
Macedo, Daniel F. ;
Nogueira, Jose Marcos S. ;
Langar, Rami ;
Secci, Stefano .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2018, 17 (11) :2660-2674
[24]   A Generalized Processor Sharing Approach to Flow Control in Integrated Services Networks: The Single-Node Case [J].
Parekh, Abhay K. ;
Gallager, Robert G. .
IEEE-ACM TRANSACTIONS ON NETWORKING, 1993, 1 (03) :344-357
[25]   Survey on Multi-Access Edge Computing for Internet of Things Realization [J].
Porambage, Pawani ;
Okwuibe, Jude ;
Liyanage, Madhusanka ;
Ylianttila, Mika ;
Taleb, Tarik .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2018, 20 (04) :2961-2991
[26]  
Poularakis K, 2019, IEEE INFOCOM SER, P10, DOI [10.1109/INFOCOM.2019.8737385, 10.1109/infocom.2019.8737385]
[27]   Hierarchical Fog-Cloud Computing for IoT Systems: A Computation Offloading Game [J].
Shah-Mansouri, Hamed ;
Wong, Vincent W. S. .
IEEE INTERNET OF THINGS JOURNAL, 2018, 5 (04) :3246-3257
[28]  
Speedtest intelligence, SPEEDT MARK REP CAN
[29]  
Van Hasselt H., 2018, ARXIV181202648V1 CSA
[30]  
van Hasselt H, 2016, AAAI CONF ARTIF INTE, P2094