Adaptive Video Streaming With Edge Caching and Video Transcoding Over Software-Defined Mobile Networks: A Deep Reinforcement Learning Approach

被引:80
作者
Luo, Jia [1 ]
Yu, F. Richard [2 ]
Chen, Qianbin [1 ]
Tang, Lun [1 ]
机构
[1] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Mobile Commun Technol, Chongqing 400065, Peoples R China
[2] Carleton Univ, Dept Syst & Comp Engn, Ottawa, ON K1S 5B6, Canada
基金
中国国家自然科学基金;
关键词
Streaming media; Quality of experience; Transcoding; Adaptation models; Bit rate; Markov processes; Cloud computing; Software defined mobile networks; mobile edge cloud; adaptive video streaming; Lyapunov technique; deep reinforcement learning; WIRELESS CELLULAR NETWORKS; RESOURCE-ALLOCATION; ADAPTATION; MANAGEMENT; TIME;
D O I
10.1109/TWC.2019.2955129
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Both mobile edge cloud (MEC) and software-defined networking (SDN) are technologies for next generation mobile networks. In this paper, we propose to simultaneously optimize energy consumption and quality of experience (QoE) metrics in video streaming over software-defined mobile networks (SDMN) combined with MEC. Specifically, we propose a novel mechanism to jointly consider buffer dynamics, video quality adaption, edge caching, video transcoding and transmission. First, we assume that the time-varying channel is a discrete-time Markov chain (DTMC). Then, based on this assumption, we formulate two optimization problems which can be depicted as a constrained Markov decision process (CMDP) and a Markov decision process (MDP). Then, we transform the CMDP problem into regular MDP by deploying Lyapunov technique. We utilize asynchronous advantage actor-critic (A3C) algorithm, one of the model-free deep reinforcement learning (DRL) methods, to solve the corresponding MDP issues. Simulation results are presented to show that the proposed scheme can achieve the goal of energy saving and QoE enhancement with the corresponding constraints satisfied.
引用
收藏
页码:1577 / 1592
页数:16
相关论文
共 41 条
[1]  
[Anonymous], IEEE T EMERG TOPICS
[2]  
[Anonymous], 1454457600805266 CIS
[3]  
[Anonymous], P IEEE GLOBECOM
[4]   Deep Reinforcement Learning A brief survey [J].
Arulkumaran, Kai ;
Deisenroth, Marc Peter ;
Brundage, Miles ;
Bharath, Anil Anthony .
IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) :26-38
[5]   Computation Rate Maximization for Wireless Powered Mobile-Edge Computing With Binary Computation Offloading [J].
Bi, Suzhi ;
Zhang, Ying Jun .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2018, 17 (06) :4177-4190
[6]   Playout Continuity Driven Framework for HTTP Adaptive Streaming Over LTE Networks [J].
Chen, Yuchen ;
Liu, Guizhong .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2018, 28 (02) :468-482
[7]   On Improving Video Streaming Efficiency, Fairness, Stability, and Convergence Time Through Client-Server Cooperation [J].
El Marai, Oussama ;
Taleb, Tarik ;
Menacer, Mohamed ;
Koudil, Mouloud .
IEEE TRANSACTIONS ON BROADCASTING, 2018, 64 (01) :11-25
[8]   On Intelligent Traffic Control for Large-Scale Heterogeneous Networks: A Value Matrix-Based Deep Learning Approach [J].
Fadlullah, Zubair Md. ;
Tang, Fengxiao ;
Mao, Bomin ;
Liu, Jiajia ;
Kato, Nei .
IEEE COMMUNICATIONS LETTERS, 2018, 22 (12) :2479-2482
[9]  
Farmanbar H, 2014, IEEE IFIP NETW OPER
[10]  
George Anna L., 2010, Southeastern Fishes Council Proceedings, V52, P1