Fast Adaptive Task Offloading in Edge Computing Based on Meta Reinforcement Learning

被引:287
作者
Wang, Jin [1 ]
Hu, Jia [1 ]
Min, Geyong [1 ]
Zomaya, Albert Y. [2 ]
Georgalas, Nektarios [3 ]
机构
[1] Univ Exeter, Dept Comp Sci, Exeter EX4 4PY, Devon, England
[2] Univ Sydney, Sch Informat Technol, Sydney, NSW 2006, Australia
[3] British Telecommun PLC, Dept Appl Res, Edinburgh EH12, Midlothian, Scotland
关键词
Task analysis; Training; Neural networks; Heuristic algorithms; Mobile applications; Learning (artificial intelligence); Edge computing; Multi-access edge computing; task offloading; meta reinforcement learning; deep learning; MOBILE; ALGORITHM;
D O I
10.1109/TPDS.2020.3014896
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Multi-access edge computing (MEC) aims to extend cloud service to the network edge to reduce network traffic and service latency. A fundamental problem in MEC is how to efficiently offload heterogeneous tasks of mobile applications from user equipment (UE) to MEC hosts. Recently, many deep reinforcement learning (DRL)-based methods have been proposed to learn offloading policies through interacting with the MEC environment that consists of UE, wireless channels, and MEC hosts. However, these methods have weak adaptability to new environments because they have low sample efficiency and need full retraining to learn updated policies for new environments. To overcome this weakness, we propose a task offloading method based on meta reinforcement learning, which can adapt fast to new environments with a small number of gradient updates and samples. We model mobile applications as Directed Acyclic Graphs (DAGs) and the offloading policy by a custom sequence-to-sequence (seq2seq) neural network. To efficiently train the seq2seq network, we propose a method that synergizes the first order approximation and clipped surrogate objective. The experimental results demonstrate that this new offloading method can reduce the latency by up to 25 percent compared to three baselines while being able to adapt fast to new environments.
引用
收藏
页码:242 / 253
页数:12
相关论文
共 40 条
[1]   Energy-Efficient Resource Allocation for Mobile Edge Computing-Based Augmented Reality Applications [J].
Al-Shuwaili, Ali ;
Simeone, Osvaldo .
IEEE WIRELESS COMMUNICATIONS LETTERS, 2017, 6 (03) :398-401
[2]  
[Anonymous], 2018, ARXIVABS181003548
[3]   List Scheduling Algorithm for Heterogeneous Systems by an Optimistic Cost Table [J].
Arabnejad, Hamid ;
Barbosa, Jorge G. .
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2014, 25 (03) :682-694
[4]  
Ba J. L., 2016, LAYER NORMALIZATION
[5]  
Bahdanau D, 2016, Arxiv, DOI [arXiv:1409.0473, DOI 10.48550/ARXIV.1409.0473]
[6]   Avatar: Mobile Distributed Computing in the Cloud [J].
Borcea, Cristian ;
Ding, Xiaoning ;
Gehani, Narain ;
Curtmola, Reza ;
Khan, Mohammad A. ;
Debnath, Hillol .
2015 3RD IEEE INTERNATIONAL CONFERENCE ON MOBILE CLOUD COMPUTING, SERVICES, AND ENGINEERING (MOBILECLOUD 2015), 2015, :151-156
[7]   Reinforcement Learning, Fast and Slow [J].
Botvinick, Matthew ;
Ritter, Sam ;
Wang, Jane X. ;
Kurth-Nelson, Zeb ;
Blundell, Charles ;
Hassabis, Demis .
TRENDS IN COGNITIVE SCIENCES, 2019, 23 (05) :408-422
[8]   Task Offloading for Mobile Edge Computing in Software Defined Ultra-Dense Network [J].
Chen, Min ;
Hao, Yixue .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2018, 36 (03) :587-597
[9]   Optimized Computation Offloading Performance in Virtual Edge Computing Systems via Deep Reinforcement Learning [J].
Chen, Xianfu ;
Zhang, Honggang ;
Wu, Celimuge ;
Mao, Shiwen ;
Ji, Yusheng ;
Bennis, Mehdi .
IEEE INTERNET OF THINGS JOURNAL, 2019, 6 (03) :4005-4018
[10]  
Chinchali S, 2018, AAAI CONF ARTIF INTE, P766