Proactive Content Caching Based on Actor-Critic Reinforcement Learning for Mobile Edge Networks

被引:16
作者
Jiang, Wei [1 ]
Feng, Daquan [1 ]
Sun, Yao [2 ]
Feng, Gang [3 ,4 ]
Wang, Zhenzhong [5 ]
Xia, Xiang-Gen [6 ]
机构
[1] Shenzhen Univ, Guangdong Prov Engn Lab Digital Creat Technol, Shenzhen Key Lab Digital Creat Technol, Coll Elect & Informat Engn,Guangdong Key Lab Inte, Shenzhen 518060, Peoples R China
[2] Univ Glasgow, James Watt Sch Engn, Glasgow G12 8QQ, Lanark, Scotland
[3] Univ Elect Sci & Technol China, Yangtze Delta Reg Inst Huzhou, Huzhou 313001, Scotland
[4] Univ Elect Sci & Technol China, Natl Key Lab Sci & Technol Commun, Chengdu 611731, Peoples R China
[5] Tech Management Ctr, China Media Grp, Beijing 100020, Peoples R China
[6] Univ Delaware, Dept Elect & Comp Engn, Newark, DE 19716 USA
基金
国家重点研发计划;
关键词
Actor-critic algorithm; branching neural network; reinforcement learning; mobile edge caching; 5G NETWORKS; SMALL-CELL; DELIVERY; POLICY;
D O I
10.1109/TCCN.2021.3130995
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Mobile edge caching/computing (MEC) has emerged as a promising approach for addressing the drastic increasing mobile data traffic by bringing high caching and computing capabilities to the edge of networks. Under MEC architecture, content providers (CPs) are allowed to lease some virtual machines (VMs) at MEC servers to proactively cache popular contents for improving users' quality of experience. The scalable cache resource model rises the challenge for determining the ideal number of leased VMs for CPs to obtain the minimum expected downloading delay of users at the lowest caching cost. To address these challenges, in this paper, we propose an actor-critic (AC) reinforcement learning based proactive caching policy for mobile edge networks without the prior knowledge of users' content demand. Specifically, we formulate the proactive caching problem under dynamical users' content demand as a Markov decision process and propose a AC based caching algorithm to minimize the caching cost and the expected downloading delay. Particularly, to reduce the computational complexity, a branching neural network is employed to approximate the policy function in the actor part. Numerical results show that the proposed caching algorithm can significantly reduce the total cost and the average downloading delay when compared with other popular algorithms.
引用
收藏
页码:1239 / 1252
页数:14
相关论文
共 50 条
  • [21] Asymmetric Actor-Critic for Adapting to Changing Environments in Reinforcement Learning
    Yue, Wangyang
    Zhou, Yuan
    Zhang, Xiaochuan
    Hua, Yuchen
    Li, Minne
    Fan, Zunlin
    Wang, Zhiyuan
    Kou, Guang
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT IV, 2024, 15019 : 325 - 339
  • [22] Dynamic Pricing Based on Demand Response Using Actor-Critic Agent Reinforcement Learning
    Ismail, Ahmed
    Baysal, Mustafa
    ENERGIES, 2023, 16 (14)
  • [23] Lexicographic Actor-Critic Deep Reinforcement Learning for Urban Autonomous Driving
    Zhang, Hengrui
    Lin, Youfang
    Han, Sheng
    Lv, Kai
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (04) : 4308 - 4319
  • [24] Influences of Reinforcement and Choice Histories on Choice Behavior in Actor-Critic Learning
    Katahira K.
    Kimura K.
    Computational Brain & Behavior, 2023, 6 (2) : 172 - 194
  • [25] An Actor-Critic Reinforcement Learning Approach for Energy Harvesting Communications Systems
    Masadeh, Ala'eddin
    Wang, Zhengdao
    Kamal, Ahmed E.
    2019 28TH INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATION AND NETWORKS (ICCCN), 2019,
  • [26] On the sample complexity of actor-critic method for reinforcement learning with function approximation
    Kumar, Harshat
    Koppel, Alec
    Ribeiro, Alejandro
    MACHINE LEARNING, 2023, 112 (07) : 2433 - 2467
  • [27] Automated State Feature Learning for Actor-Critic Reinforcement Learning through NEAT
    Peng, Yiming
    Chen, Gang
    Holdaway, Scott
    Mei, Yi
    Zhang, Mengjie
    PROCEEDINGS OF THE 2017 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION (GECCO'17 COMPANION), 2017, : 135 - 136
  • [28] Actor-Critic Reinforcement Learning for Automatic Left Atrial Appendage Segmentation
    Abdullah, Al Walid
    Yun, Il Dong
    PROCEEDINGS 2018 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE (BIBM), 2018, : 609 - 612
  • [29] Supervised actor-critic reinforcement learning with action feedback for algorithmic trading
    Qizhou Sun
    Yain-Whar Si
    Applied Intelligence, 2023, 53 : 16875 - 16892
  • [30] On the sample complexity of actor-critic method for reinforcement learning with function approximation
    Harshat Kumar
    Alec Koppel
    Alejandro Ribeiro
    Machine Learning, 2023, 112 : 2433 - 2467