Resource Management for Power-Constrained HEVC Transcoding Using Reinforcement Learning

被引:9
作者
Costero, Luis [1 ]
Iranfar, Arman [2 ]
Zapater, Marina [2 ,3 ]
Igual, Francisco D. [1 ]
Olcoz, Katzalin [1 ]
Atienza, David [2 ]
机构
[1] Univ Complutense Madrid, Dept Arquitectura Comp & Automat, Madrid 28040, Spain
[2] Swiss Fed Inst Technol Lausanne EPFL, Embedded Syst Lab ESL, CH-1015 Lausanne, Switzerland
[3] Univ Appl Sci Western, Sch Engn & Management Vaud HEIG VD, CH-2800 Delemont, Switzerland
基金
欧盟地平线“2020”;
关键词
Resource management; DVFS; power capping; reinforcement learning; Q-learning; HEVC; self-adaptation;
D O I
10.1109/TPDS.2020.3004735
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
The advent of online video streaming applications and services along with the users' demand for high-quality contents require High Efficiency Video Coding (HEVC), which provides higher video quality and more compression at the cost of increased complexity. On one hand, HEVC exposes a set of dynamically tunable parameters to provide trade-offs among Quality-of-Service (QoS), performance, and power consumption of multi-core servers on the video providers' data center. On the other hand, resource management of modern multi-core servers is in charge of adapting system-level parameters, such as operating frequency and multithreading, to deal with concurrent applications and their requirements. Therefore, efficient multi-user HEVC streaming necessitates joint adaptation of application- and system-level parameters. Nonetheless, dealing with such a large and dynamic design space is challenging and difficult to address through conventional resource management strategies. Thus, in this work, we develop a multi-agent Reinforcement Learning framework to jointly adjust application- and system-level parameters at runtime to satisfy the QoS of multi-user HEVC streaming in power-constrained servers. In particular, the design space, composed of all design parameters, is split into smaller independent sub-spaces. Each design sub-space is assigned to a particular agent so that it can explore it faster, yet accurately. The benefits of our approach are revealed in terms of adaptability and quality (with up to to 4x improvements in terms of QoS when compared to a static resource management scheme), and learning time (6 x faster than an equivalent mono-agent implementation). Finally, we show that the power-capping techniques formulated outperform the hardware-based power capping with respect to quality.
引用
收藏
页码:2834 / 2850
页数:17
相关论文
共 41 条
[1]  
Ahn Y.-J., 2013, VISUAL COMMUN-US, P1
[2]  
[Anonymous], 2013, JCTVCL1100
[3]  
[Anonymous], 1998, REINFORCEMENT LEARNI
[4]   Applying reinforcement learning towards automating resource allocation and application scalability in the cloud [J].
Barrett, Enda ;
Howley, Enda ;
Duggan, Jim .
CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2013, 25 (12) :1656-1674
[5]  
Biatek T, 2014, EUR SIGNAL PR CONF, P296
[6]   Evaluation of DVFS techniques on modern HPC processors and accelerators for energy-aware applications [J].
Calore, Enrico ;
Gabbana, Alessandro ;
Schifano, Sebastiano Fabio ;
Tripiccione, Raffaele .
CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2017, 29 (12)
[7]  
Costero L, 2019, DES AUT TEST EUROPE, P558, DOI [10.23919/date.2019.8715256, 10.23919/DATE.2019.8715256]
[8]  
Das Abhishek., 2014, Electrical Engineering Congress (iEECON), 2014 International, P1, DOI DOI 10.7873/DATE.2014.115
[9]   SOSA: Self-Optimizing Learning with Self-Adaptive Control for Hierarchical System-on-Chip Management [J].
Donyanavard, Bryan ;
Muck, Tiago ;
Rahmani, Amir M. ;
Dutt, Nikil ;
Sadighi, Armin ;
Maurer, Florian ;
Herkersdorf, Andreas .
MICRO'52: THE 52ND ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE, 2019, :685-698
[10]  
Even-Dar E, 2003, J MACH LEARN RES, V5, P1