Deep Reinforcement Learning based Service Migration Strategy for Edge Computing

被引:38
作者
Gao, Zhipeng [1 ]
Jiao, Qidong [1 ]
Xiao, Kaile [1 ]
Wang, Qian [1 ]
Mo, Zijia [1 ]
Yang, Yang [1 ]
机构
[1] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing 100876, Peoples R China
来源
2019 13TH IEEE INTERNATIONAL CONFERENCE ON SERVICE-ORIENTED SYSTEM ENGINEERING (SOSE) / 10TH INTERNATIONAL WORKSHOP ON JOINT CLOUD COMPUTING (JCC) / IEEE INTERNATIONAL WORKSHOP ON CLOUD COMPUTING IN ROBOTIC SYSTEMS (CCRS) | 2019年
关键词
Edge Computing; service migration; reinforcement learning; Deep Q Network; cost;
D O I
10.1109/SOSE.2019.00025
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Edge Computing (EC) is an emerging technology to cope with the unprecedented growth of user demands for access to low-latency computation and content data. However, user mobility and limited coverage of Edge Computing Server (ECS) result in service discontinuity and reduce Quality of Service (QoS). Service migration has a great potential to address this issue. In the scenario of service migration, how to choose the optimal migration strategy and communication strategy is a key challenge. In this paper, we innovatively propose solving the service migration using reinforcement learning based model which can take a long-term goal into consideration and make service migration and communication decisions more efficient. we consider a single-user EC system with exploiting predefined movement of user, where user passes through many ECSs and its corresponding Virtual Machine (VM) in ECS decides the migration strategy and communication strategy. We design a Reinforcement Learning (RL)-based framework for a single-user EC service migration system. Q-learning based and Deep Q Network (DQN) based themes are analyzed in detail respectively. Simulation results shows that our RL-based system can achieve the optimal result compared with other two methods under different system parameters.
引用
收藏
页码:116 / 121
页数:6
相关论文
共 19 条
[1]  
Ibrahim M., 2009, INT S MOD OPT MOB AD
[2]   Reinforcement learning: A survey [J].
Kaelbling, LP ;
Littman, ML ;
Moore, AW .
JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 1996, 4 :237-285
[3]  
Ksentini A., 2014, IEEE INT C COMM ICC
[4]  
Kuang KX, 2014, IEEE INT CONF HIGH, P1, DOI 10.1109/HPSR.2014.6900873
[5]  
Lobillo F, 2014, IEEE WIREL COMMUNN, P1, DOI 10.1109/WCNCW.2014.6934851
[6]   Mobile Edge Computing: A Survey on Architecture and Computation Offloading [J].
Mach, Pavel ;
Becvar, Zdenek .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2017, 19 (03) :1628-1656
[7]   A Destination and Mobility Path Prediction Scheme for Mobile Networks [J].
Nadembega, Apollinaire ;
Hafid, Abdelhakim ;
Taleb, Tarik .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2015, 64 (06) :2577-2590
[8]  
Plachy J, 2016, 2016 IEEE 27TH ANNUAL INTERNATIONAL SYMPOSIUM ON PERSONAL, INDOOR, AND MOBILE RADIO COMMUNICATIONS (PIMRC), P2384
[9]   Path selection enabling user mobility and efficient distribution of data for computation at the edge of mobile network [J].
Plachy, Jan ;
Becvar, Zdenek ;
Mach, Pavel .
COMPUTER NETWORKS, 2016, 108 :357-370
[10]  
Sadilek A., 2012, FAR OUT PREDICTING L