Online Learning for Offloading and Autoscaling in Energy Harvesting Mobile Edge Computing

被引:317
作者
Xu, Jie [1 ]
Chen, Lixing [1 ]
Ren, Shaolei [2 ]
机构
[1] Univ Miami, Dept Elect & Comp Engn, Miami, FL 33146 USA
[2] Univ Calif Riverside, Dept Elect & Comp Engn, Riverside, CA 92521 USA
基金
美国国家科学基金会;
关键词
Mobile edge computing; energy harvesting; online learning; MANAGEMENT;
D O I
10.1109/TCCN.2017.2725277
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Mobile edge computing (also known as fog computing) has recently emerged to enable in-situ processing of delay-sensitive applications at the edge of mobile networks. Providing grid power supply in support of mobile edge computing, however, is costly and even infeasible (in certain rugged or under-developed areas), thus mandating on-site renewable energy as a major or even sole power supply in increasingly many scenarios. Nonetheless, the high intermittency and unpredictability of renewable energy make it very challenging to deliver a high quality of service to users in energy harvesting mobile edge computing systems. In this paper, we address the challenge of incorporating renewables into mobile edge computing and propose an efficient reinforcement learning-based resource management algorithm, which learns on-the-fly the optimal policy of dynamic workload offloading (to the centralized cloud) and edge server provisioning to minimize the long-term system cost (including both service delay and operational cost). Our online learning algorithm uses a decomposition of the (offline) value iteration and (online) reinforcement learning, thus achieving a significant improvement of learning rate and run-time performance when compared to standard reinforcement learning algorithms such as Q-learning. We prove the convergence of the proposed algorithm and analytically show that the learned policy has a simple monotone structure amenable to practical implementation. Our simulation results validate the efficacy of our algorithm, which significantly improves the edge computing performance compared to fixed or myopic optimization schemes and conventional reinforcement learning algorithms.
引用
收藏
页码:361 / 373
页数:13
相关论文
共 30 条
[1]  
[Anonymous], PROC IEEE GLOB COMM
[2]  
[Anonymous], 2014, Markov decision processes: discrete stochastic dynamic programming
[3]  
[Anonymous], 2011, INT C HIGH PERFORMAN, DOI DOI 10.1145/2063384.2063411
[4]  
Beck MT., 2014, Proceedings of the 8th International Conference on Advanced Engineering Computing and Applications in Sciences (ADVCOMP), P65
[5]  
Bertsekas DP, 1995, Dynamic programming and optimal control
[6]   The ODE method for convergence of stochastic approximation and reinforcement learning [J].
Borkar, VS ;
Meyn, SP .
SIAM JOURNAL ON CONTROL AND OPTIMIZATION, 2000, 38 (02) :447-469
[7]  
Chen LB, 2017, IEEE INT SYMP NANO, P1, DOI 10.1109/NANOARCH.2017.8053709
[8]   Fog and IoT: An Overview of Research Opportunities [J].
Chiang, Mung ;
Zhang, Tao .
IEEE INTERNET OF THINGS JOURNAL, 2016, 3 (06) :854-864
[9]   Towards Power Consumption-Delay Tradeoff by Workload Allocation in Cloud-Fog Computing [J].
Deng, Ruilong ;
Lu, Rongxing ;
Lai, Chengzhe ;
Luan, Tom H. .
2015 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2015, :3909-3914
[10]   Performance Analysis of Wireless Energy Harvesting Cognitive Radio Networks under Smart Jamming Attacks [J].
Hoang, Dinh Thai ;
Niyato, Dusit ;
Wang, Ping ;
Kim, Dong In .
IEEE Transactions on Cognitive Communications and Networking, 2015, 1 (02) :200-216