Attention Mechanism-Aided Deep Reinforcement Learning for Dynamic Edge Caching

被引:3
作者
Teng, Ziyi [1 ]
Fang, Juan [1 ]
Yang, Huijing [1 ]
Yu, Lu [2 ]
Chen, Huijie [1 ]
Xiang, Wei [3 ]
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing 100124, Peoples R China
[2] James Cook Univ, Dept Elect & Comp Engn, Cairns, Qld 4878, Australia
[3] La Trobe Univ, Sch Comp Engn & Math Sci, Melbourne, Vic 3086, Australia
基金
中国国家自然科学基金;
关键词
Servers; Wireless communication; Optimization; Load modeling; Resource management; Internet of Things; Telecommunication traffic; Attention-weighted channel assignment; deep reinforcement learning; edge caching; wireless network; USER ASSOCIATION; PLACEMENT;
D O I
10.1109/JIOT.2023.3327656
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The dynamic mechanism of joint proactive caching and cache replacement, which involves placing content items close to cache-enabled edge devices ahead of time until they are requested, is a promising technique for enhancing traffic offloading and relieving heavy network loads. However, due to limited edge cache capacity and wireless transmission resources, accurately predicting users' future requests and performing dynamic caching is crucial to effectively utilizing these limited resources. This article investigates joint proactive caching and cache replacement strategies in a general mobile-edge computing (MEC) network with multiple users under a cloud-edge-device collaboration architecture. The joint optimization problem is formulated as a Markov decision process (MDP) problem with an infinite range of average network load costs, aiming to reduce network load traffic while efficiently utilizing the limited available transport resources. To address this issue, we design an attention-weighted deep deterministic policy gradient (AWD2PG) model, which uses attention weights to allocate the number of channels from server to user, and applies deep deterministic policies on both user and server sides for Cache decision-making, so as to achieve the purpose of reducing network traffic load and improving network and cache resource utilization. We verify the convergence of the corresponding algorithms and demonstrate the effectiveness of the proposed AWD2PG strategy and benchmark in reducing network load and improving hit rate.
引用
收藏
页码:10197 / 10213
页数:17
相关论文
共 31 条
[1]   A proactive caching and offloading technique using machine learning for mobile edge computing users [J].
Alqahtani, Fayez ;
Al-Maitah, Mohammed ;
Elshakankiry, Osama .
COMPUTER COMMUNICATIONS, 2022, 181 :224-235
[2]   Cache-Enabled Multicast Content Pushing With Structured Deep Learning [J].
Chen, Qi ;
Wang, Wei ;
Chen, Wei ;
Yu, F. Richard ;
Zhang, Zhaoyang .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2021, 39 (07) :2135-2149
[3]   Content Pushing With Request Delay Information [J].
Chen, Wei ;
Poor, H. Vincent .
IEEE TRANSACTIONS ON COMMUNICATIONS, 2017, 65 (03) :1146-1161
[4]   IoT Application Modules Placement and Dynamic Task Processing in Edge-Cloud Computing [J].
Fang, Juan ;
Ma, Aonan .
IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (16) :12771-12781
[5]  
Fujimoto S, 2018, PR MACH LEARN RES, V80
[6]   Reinforcement Learning Based Cooperative Coded Caching Under Dynamic Popularities in Ultra-Dense Networks [J].
Gao, Shen ;
Dong, Peihao ;
Pan, Zhiwen ;
Li, Geoffrey Ye .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (05) :5442-5456
[7]   Wireless Content Caching for Small Cell and D2D Networks [J].
Gregori, Maria ;
Gomez-Vilardebo, Jesus ;
Matamoros, Javier ;
Gunduz, Deniz .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2016, 34 (05) :1222-1234
[8]  
Gummadi KrishnaP., 2003, Proceedings of the Nineteenth ACM Symposium on Operating Systems Principles, SOSP '03, P314
[9]  
Hachem Jad, 2015, 2015 IEEE Conference on Computer Communications (INFOCOM). Proceedings, P756, DOI 10.1109/INFOCOM.2015.7218445
[10]  
Lillicrap T.P., 2015, arXiv, DOI [10.48550/arXiv.1509.02971, DOI 10.48550/ARXIV.1509.02971]