SACC: A Size Adaptive Content Caching Algorithm in Fog/Edge Computing Using Deep Reinforcement Learning

被引:11
作者
Zhou, Xiaoping [1 ,2 ]
Liu, Zhenlong [3 ]
Guo, Maozu [1 ]
Zhao, Jichao [4 ]
Wang, Jialin [5 ]
机构
[1] Beijing Univ Civil Engn & Architecture, Beijing Key Lab Intelligent Proc Bldg Big Data, Beijing 100044, Peoples R China
[2] Natl Univ Singapore, Singapore 119077, Singapore
[3] Beijing Univ Posts & Telecommun, Sch Cyberspace Secur, Beijing 100876, Peoples R China
[4] CNPC Managers Training Inst, Beijing 100007, Peoples R China
[5] BOSWinner Co Ltd, Smart Spaces Res Inst, Beijing 315803, Peoples R China
基金
北京市自然科学基金;
关键词
Special-purpose and application-based systems; machine learning; content caching; deep reinforcement learning; size adaptive; actor-critic framework;
D O I
10.1109/TETC.2021.3115793
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Edge/Fog caching is promising to mitigate the data traffic problem in both traditional wireline/wireless networks and the 5G network. Recently, deep reinforcement learning (DRL) has been adopted to provide a more powerful content caching policy. The current DRL-based scheme considers the requests for the same size and updates the caching for each request. However, the real-world data delivery systems usually refresh the content cache periodically, with different sizes of requests. To satisfy the real-world requirements, this study proposes a novel size adaptive content caching algorithm using DRL, termed SACC. SACC models the requests with random sizes and updates the cache after a batch of requests. Technically, SACC utilizes the Actor-Critic framework, which is able to process large discrete action space. SACC comprehensively considers the short-, medium- and long-term requests as the state to train the actor network. The reward is modeled as the cache hit rate. Once an action is selected from the policy network, it is expended to its k nearest neighbors. The critic network finds the action with the best reward from the k actions. The performance of the proposed SACC is evaluated through computer simulation. The experimental results showed that SACC could train the network much more efficiently and improve the cache hit rate by as much as 4% when comparing to the state-of-art DRL-based scheme.
引用
收藏
页码:1810 / 1820
页数:11
相关论文
共 36 条
[1]  
Ahmed M, 2013, Arxiv, DOI arXiv:1301.4909
[2]  
[Anonymous], 2011, Relatorio tecnico, Telecom ParisTech 2011
[3]  
Blasco P, 2014, IEEE ICC, P1897, DOI 10.1109/ICC.2014.6883600
[4]  
C. V. N. Index, 2019, CISCO VISUAL NETWORK
[5]   Co-Operative Caching in Dynamic Shared Spectrum Networks [J].
Das, Dibakar ;
Abouzeid, Alhussein A. .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2016, 15 (07) :5060-5075
[6]  
Dulac-Arnold G, 2016, Arxiv, DOI arXiv:1512.07679
[7]   Performance evaluation of the random replacement policy for networks of caches [J].
Gallo, Massimo ;
Kauffmann, Bruno ;
Muscariello, Luca ;
Simonian, Alain ;
Tanguy, Christian .
PERFORMANCE EVALUATION, 2014, 72 :16-36
[8]   A Survey of Actor-Critic Reinforcement Learning: Standard and Natural Policy Gradients [J].
Grondman, Ivo ;
Busoniu, Lucian ;
Lopes, Gabriel A. D. ;
Babuska, Robert .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2012, 42 (06) :1291-1307
[9]  
Guan Y, 2014, IEEE GLOB COMM CONF, P4537, DOI 10.1109/GLOCOM.2014.7037523
[10]   Joint Switch Upgrade and Controller Deployment in Hybrid Software-Defined Networks [J].
Guo, Zehua ;
Chen, Weikun ;
Liu, Ya-Feng ;
Xu, Yang ;
Zhang, Zhi-Li .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2019, 37 (05) :1012-1028