DERP: A Deep Reinforcement Learning Cloud System for Elastic Resource Provisioning

被引:41
作者
Bitsakos, Constantinos [1 ]
Konstantinou, Ioannis [1 ]
Koziris, Nectarios [1 ]
机构
[1] Natl Tech Univ Athens, Sch Elect & Comp Engn, Athens, Greece
来源
2018 16TH IEEE INTERNATIONAL CONFERENCE ON CLOUD COMPUTING TECHNOLOGY AND SCIENCE (CLOUDCOM 2018) | 2018年
基金
欧盟地平线“2020”;
关键词
Elasticity; Resource Management; Resource Provisioning; Cloud computing; Deep Reinforecement learning; Double deep Q learning; NoSQL databases; DERP;
D O I
10.1109/CloudCom2018.2018.00020
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Modern large scale computer clusters benefit significantly from elasticity. Elasticity allows a cluster to dynamically allocate computer resources, based on the user's fluctuating workload demands. Many cloud providers use threshold-based approaches, which have been proven to be difficult to configure and optimise, while others use reinforcement learning and decision-tree approaches, which struggle when having to handle large multidimensional cluster states. In this work we use Deep Reinforcement learning techniques to achieve automatic elasticity. We use three different approaches of a Deep Reinforcement learning agent, called DERP (Deep Elastic Resource Provisioning), that takes as input the current multi-dimensional state of a cluster and manages to train and converge to the optimal elasticity behaviour after a finite amount of training steps. The system automatically decides and proceeds on requesting/releasing VM resources from the provider and orchestrating them inside a NoSQL cluster according to user-defined policies/rewards. We compare our agent to state-of-the-art, Reinforcement learning and decision-tree based, approaches in demanding simulation environments and show that it gains rewards up to 1.6 times better on its lifetime. We then test our approach in a real life cluster environment and show that the system resizes clusters in real-time and adapts its performance through a variety of demanding optimisation strategies, input and training loads.
引用
收藏
页码:21 / 29
页数:9
相关论文
共 19 条
[1]  
[Anonymous], P ADV NEUR INF PROC
[2]  
[Anonymous], 2013, Playing atari with deep reinforcement learning
[3]  
[Anonymous], 1985, CALIFORNIA U SAN DIE
[4]  
[Anonymous], 2010, P 1 ACM S CLOUD COMP, DOI DOI 10.1145/1807128.1807152
[5]  
[Anonymous], 2009, NATL I STAND TECHNOL, DOI DOI 10.6028/NIST.SP.800-145
[6]  
Giannakopoulos Ioannis, 2014, 2014 IEEE International Conference on Big Data (Big Data), P23, DOI 10.1109/BigData.2014.7004481
[7]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[8]  
Lolos K, 2017, IEEE INT CONF BIG DA, P203, DOI 10.1109/BigData.2017.8257928
[9]   The ganglia distributed monitoring system: design, implementation, and experience [J].
Massie, ML ;
Chun, BN ;
Culler, DE .
PARALLEL COMPUTING, 2004, 30 (07) :817-840
[10]  
Nguyen H., 2013, P 10 INT C AUTONOMIC, P69