Deep reinforcement learning-based incentive mechanism design for short video sharing through D2D communication

被引:3
作者
Li, Zhuo [1 ]
Dong, Wentao [1 ]
Chen, Xin [1 ]
机构
[1] Beijing Informat Sci & Technol Univ, Sch Comp Sci, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Short video sharing; D2D communications; Dynamic incentive mechanism; Deep reinforcement learning; EDGE; 5G; CHALLENGES; BLOCKCHAIN;
D O I
10.1007/s12083-021-01146-x
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the development of 5th generation (5G) wireless communication networks and the popularity of short video applications, there has been a rapid increase in short video traffic in cellular networks. Device-to-device (D2D) communication-based short video sharing is considered to be an effective way to offload traffic from cellular networks. Due to the selfish nature of mobile user equipment (MUEs), how to dynamically motivate MUEs to engage in short video sharing while ensuring the Quality of Service, which makes it critical to design an appropriate incentive mechanism. In this paper, we firstly analyze the rationale for dynamically setting rewards and penalties and then define the rewards and penalties setting dynamically for maximizing the utility of the mobile edge computing server (RPSDMU) problem. The problem is proved NP-hard. Furthermore, we formulate the dynamic incentive process as the Markov Decision Process problem. Considering the complexity and dynamics of the problem, we design a Dynamic Incentive Mechanism algorithm of D2D-based Short Video Sharing based on Asynchronous Advantage Actor-Critic (DIM-A3C) to solve the problem. Simulation results show that the proposed dynamic incentive mechanism can increase the utility of mobile edge computing server by an average of 22% and 16% compared with the existing proportional incentive mechanism (PIM) and scoring-based incentive mechanism (SIM). Meanwhile, DIM-A3C achieves a higher degree of satisfaction than PIM and SIM.
引用
收藏
页码:3946 / 3958
页数:13
相关论文
共 39 条
[1]   QoE awareness in progressive caching and DASH-based D2D video streaming in cellular networks [J].
Al-Habashna, Ala'a ;
Wainer, Gabriel .
WIRELESS NETWORKS, 2020, 26 (03) :2051-2073
[2]   A Survey on Device-to-Device Communication in Cellular Networks [J].
Asadi, Arash ;
Wang, Qing ;
Mancuso, Vincenzo .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2014, 16 (04) :1801-1819
[3]   Living on the Edge: The Role of Proactive Caching in 5G Wireless Networks [J].
Bastug, Ejder ;
Bennis, Mehdi ;
Debbah, Merouane .
IEEE COMMUNICATIONS MAGAZINE, 2014, 52 (08) :82-89
[4]   A Coordinated Collision Mitigation Approach for Virtual Coupling Trains by Using Model Predictive Control [J].
Chen, Mingliang ;
Xun, Jing ;
Liu, Yafei .
2020 IEEE 23RD INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2020,
[5]  
Chen Z., 2016, 2016 IEEE 17th international workshop on signal processing advances in wireless communications (SPAWC), P1
[6]   Caching Incentive Design in Wireless D2D Networks: A Stackelberg Game Approach [J].
Chen, Zhuoqun ;
Liu, Yangyang ;
Zhou, Bo ;
Tao, Meixia .
2016 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2016,
[7]   Caching in Information-Centric Networking: Strategies, Challenges, and Future Research Directions [J].
Din, Ikram Ud ;
Hassan, Suhaidi ;
Khan, Muhammad Khurram ;
Guizani, Mohsen ;
Ghazali, Osman ;
Habbal, Adib .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2018, 20 (02) :1443-1474
[8]   Intelligent VNF Orchestration and Flow Scheduling via Model-Assisted Deep Reinforcement Learning [J].
Gu, Lin ;
Zeng, Deze ;
Li, Wei ;
Guo, Song ;
Zomaya, Albert Y. ;
Jin, Hai .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2020, 38 (02) :279-291
[9]   Challenges in 5G: How to Empower SON with Big Data for Enabling 5G [J].
Imran, Ali ;
Zoha, Ahmed ;
Abu-Dayya, Adnan .
IEEE NETWORK, 2014, 28 (06) :27-33
[10]   Maximized Cellular Traffic Offloading via Device-to-Device Content Sharing [J].
Jiang, Jingjie ;
Zhang, Shengkai ;
Li, Bo ;
Li, Baochun .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2016, 34 (01) :82-91