Proactive Content Caching Based on Actor-Critic Reinforcement Learning for Mobile Edge Networks

被引:16
作者
Jiang, Wei [1 ]
Feng, Daquan [1 ]
Sun, Yao [2 ]
Feng, Gang [3 ,4 ]
Wang, Zhenzhong [5 ]
Xia, Xiang-Gen [6 ]
机构
[1] Shenzhen Univ, Guangdong Prov Engn Lab Digital Creat Technol, Shenzhen Key Lab Digital Creat Technol, Coll Elect & Informat Engn,Guangdong Key Lab Inte, Shenzhen 518060, Peoples R China
[2] Univ Glasgow, James Watt Sch Engn, Glasgow G12 8QQ, Lanark, Scotland
[3] Univ Elect Sci & Technol China, Yangtze Delta Reg Inst Huzhou, Huzhou 313001, Scotland
[4] Univ Elect Sci & Technol China, Natl Key Lab Sci & Technol Commun, Chengdu 611731, Peoples R China
[5] Tech Management Ctr, China Media Grp, Beijing 100020, Peoples R China
[6] Univ Delaware, Dept Elect & Comp Engn, Newark, DE 19716 USA
基金
国家重点研发计划;
关键词
Actor-critic algorithm; branching neural network; reinforcement learning; mobile edge caching; 5G NETWORKS; SMALL-CELL; DELIVERY; POLICY;
D O I
10.1109/TCCN.2021.3130995
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Mobile edge caching/computing (MEC) has emerged as a promising approach for addressing the drastic increasing mobile data traffic by bringing high caching and computing capabilities to the edge of networks. Under MEC architecture, content providers (CPs) are allowed to lease some virtual machines (VMs) at MEC servers to proactively cache popular contents for improving users' quality of experience. The scalable cache resource model rises the challenge for determining the ideal number of leased VMs for CPs to obtain the minimum expected downloading delay of users at the lowest caching cost. To address these challenges, in this paper, we propose an actor-critic (AC) reinforcement learning based proactive caching policy for mobile edge networks without the prior knowledge of users' content demand. Specifically, we formulate the proactive caching problem under dynamical users' content demand as a Markov decision process and propose a AC based caching algorithm to minimize the caching cost and the expected downloading delay. Particularly, to reduce the computational complexity, a branching neural network is employed to approximate the policy function in the actor part. Numerical results show that the proposed caching algorithm can significantly reduce the total cost and the average downloading delay when compared with other popular algorithms.
引用
收藏
页码:1239 / 1252
页数:14
相关论文
共 50 条
  • [41] Dynamic spectrum access and sharing through actor-critic deep reinforcement learning
    Liang Dong
    Yuchen Qian
    Yuan Xing
    EURASIP Journal on Wireless Communications and Networking, 2022
  • [42] Development and Validation of Active Roll Control based on Actor-critic Neural Network Reinforcement Learning
    Bahr, Matthias
    Reicherts, Sebastian
    Sieberg, Philipp
    Morss, Luca
    Schramm, Dieter
    SIMULTECH: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON SIMULATION AND MODELING METHODOLOGIES, TECHNOLOGIES AND APPLICATIONS, 2019, 2019, : 36 - 46
  • [43] Dynamic Actor-critic: Reinforcement Learning based Radio Resource Scheduling For LTE-Advanced
    Tathe, Pallavi K.
    Sharma, Manish
    2018 FOURTH INTERNATIONAL CONFERENCE ON COMPUTING COMMUNICATION CONTROL AND AUTOMATION (ICCUBEA), 2018,
  • [44] SAC-FACT: Soft Actor-Critic Reinforcement Learning for Counterfactual Explanations
    Ezzeddine, Fatima
    Ayoub, Omran
    Andreoletti, Davide
    Giordano, Silvia
    EXPLAINABLE ARTIFICIAL INTELLIGENCE, XAI 2023, PT I, 2023, 1901 : 195 - 216
  • [45] Actor-Critic for Multi-Agent Reinforcement Learning with Self-Attention
    Zhao, Juan
    Zhu, Tong
    Xiao, Shuo
    Gao, Zongqian
    Sun, Hao
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2022, 36 (09)
  • [46] IMPROVING ACTOR-CRITIC REINFORCEMENT LEARNING VIA HAMILTONIAN MONTE CARLO METHOD
    Xu, Duo
    Fekri, Faramarz
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4018 - 4022
  • [47] CONTROLLED SENSING AND ANOMALY DETECTION VIA SOFT ACTOR-CRITIC REINFORCEMENT LEARNING
    Zhong, Chen
    Gursoy, M. Cenk
    Velipasalar, Senem
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4198 - 4202
  • [48] Towards optimal control of HPV model using safe reinforcement learning with actor-critic neural networks
    Amirabadi, Roya Khalili
    Fard, Omid S.
    Farimani, Mohsen Jalaeian
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 264
  • [49] SOFT ACTOR-CRITIC REINFORCEMENT LEARNING FOR ROBOTIC MANIPULATOR WITH HINDSIGHT EXPERIENCE REPLAY
    Yan, Tao
    Zhang, Wenan
    Yang, Simon X.
    Yu, Li
    INTERNATIONAL JOURNAL OF ROBOTICS & AUTOMATION, 2019, 34 (05) : 536 - 543
  • [50] MULTI-STEP ACTOR-CRITIC FRAMEWORK FOR REINFORCEMENT LEARNING IN CONTINUOUS CONTROL
    Huang T.
    Chen G.
    Journal of Applied and Numerical Optimization, 2023, 5 (02): : 189 - 200