Energy-Efficient Content Fetching Strategies in Cache-Enabled D2D Networks via an Actor-Critic Reinforcement Learning Structure

被引:6
作者
Yan, Ming [1 ,2 ]
Luo, Meiqi [1 ,2 ]
Chan, Chien Aun [3 ]
Gygax, Andre F. [4 ]
Li, Chunguo [5 ]
Chih-Lin, I [6 ]
机构
[1] Commun Univ China, Sch Informat & Commun Engn, Beijing 100024, Peoples R China
[2] Commun Univ China, Key Lab Acoust Visual Technol & Intelligent Contro, Beijing 100024, Peoples R China
[3] Univ Melbourne, Dept Elect & Elect Engn, Melbourne, Vic 3010, Australia
[4] Univ Melbourne, Fac Business & Econ, Melbourne, Vic 3010, Australia
[5] Southeast Univ, Sch Informat Sci & Engn, Nanjing 210096, Peoples R China
[6] China Mobile Res Inst, Beijing 100053, Peoples R China
基金
中国国家自然科学基金;
关键词
Device-to-device communication; Resource management; Energy consumption; Base stations; Optimization; Delays; Throughput; Device-to-device (D2D) networks; actor-critic algorithm; deep reinforcement learning (DRL); edge caching; fetching strategy; DEVICE CONTENT DELIVERY; OPTIMIZATION;
D O I
10.1109/TVT.2024.3419012
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
As one of the important complementary technologies of the fifth-generation (5G) wireless communication and beyond, mobile device-to-device (D2D) edge caching and computing can effectively reduce the pressure on backbone networks and improve the user experience. Specific content can be pre-cached on the user devices based on personalized content placement strategies, and the cached content can be fetched by neighboring devices in the same D2D network. However, when multiple devices simultaneously fetch content from the same device, collisions will occur and reduce communication efficiency. In this paper, we design the content fetching strategies based on an actor-critic deep reinforcement learning (DRL) architecture, which can adjust the content fetching collision rate to adapt to different application scenarios. First, the optimization problem is formulated with the goal of minimizing the collision rate to improve the throughput, and a general actor-critic DRL algorithm is used to improve the content fetching strategy. Second, by optimizing the network architecture and reward function, the two-level actor-critic algorithm is improved to effectively manage the collision rate and transmission power. Furthermore, to balance the conflict between the collision rate and device energy consumption, the related reward values are weighted in the reward function to optimize the energy efficiency. The simulation results show that the content fetching collision rate based on the improved two-level actor-critic algorithm decreases significantly compared with that of the baseline algorithms, and the network energy consumption can be optimized by adjusting the weight factors.
引用
收藏
页码:17485 / 17495
页数:11
相关论文
共 41 条
[1]   A Joint Reinforcement-Learning Enabled Caching and Cross-Layer Network Code in F-RAN With D2D Communications [J].
Al-Abiad, Mohammed S. ;
Hassan, Md Zoheb ;
Hossain, Md Jahangir .
IEEE TRANSACTIONS ON COMMUNICATIONS, 2022, 70 (07) :4400-4416
[2]   UAV Assisted Cooperative Caching on Network Edge Using Multi-Agent Actor-Critic Reinforcement Learning [J].
Araf, Sadman ;
Saha, Adittya Soukarjya ;
Kazi, Sadia Hamid ;
Tran, Nguyen H. H. ;
Alam, Md. Golam Rabiul .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (02) :2322-2337
[3]   A Deep-Reinforcement-Learning-Based Social-Aware Cooperative Caching Scheme in D2D Communication Networks [J].
Bai, Yalu ;
Wang, Dan ;
Huang, Gang ;
Song, Bin .
IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (11) :9634-9645
[4]   Dynamic Video Delivery using Deep Reinforcement Learning for Device-to-Device Underlaid Cache-Enabled Internet-of-vehicle Networks [J].
Choi, Minseok ;
Shin, Myungjae ;
Kim, Joongheon .
JOURNAL OF COMMUNICATIONS AND NETWORKS, 2021, 23 (02) :117-128
[5]  
Fan Jiang, 2021, 2021 IEEE/CIC International Conference on Communications in China (ICCC), P922, DOI 10.1109/ICCC52777.2021.9580239
[6]   Multi-Agent Collaborative Caching Strategies in Dynamic Heterogeneous D2D Networks [J].
Fan, Xinglong ;
Chen, Honglong ;
Ni, Zhichen ;
Li, Guoxin ;
Sun, Haiyang ;
Yu, Jiguo .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2024, 73 (05) :7204-7217
[7]   Caching Efficiency Maximization for Device-to-Device Communication Networks: A Recommend to Cache Approach [J].
Fu, Yaru ;
Salaun, Lou ;
Yang, Xiaolong ;
Wen, Wanli ;
Quek, Tony Q. S. .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (10) :6580-6594
[8]   Energy-Efficient Joint Task Assignment and Power Control in Energy-Harvesting D2D Offloading Communications [J].
Hamdi, Monia ;
Ben Hamed, Aws ;
Yuan, Di ;
Zaied, Mourad .
IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (08) :6018-6031
[9]   Intimacy-based Resource Allocation for Network Slicing in 5G via Deep Reinforcement Learning [J].
He, Nan ;
Yang, Song ;
Li, Fan ;
Chen, Xu .
IEEE NETWORK, 2021, 35 (06) :111-118
[10]   GBLinks: GNN-Based Beam Selection and Link Activation for Ultra-Dense D2D mmWave Networks [J].
He, Shiwen ;
Xiong, Shaowen ;
Zhang, Wei ;
Yang, Yiting ;
Ren, Ju ;
Huang, Yongming .
IEEE TRANSACTIONS ON COMMUNICATIONS, 2022, 70 (05) :3451-3466