Towards Accurate Prediction for High-Dimensional and Highly-Variable Cloud Workloads with Deep Learning

被引:114
作者
Chen, Zheyi [1 ]
Hu, Jia [1 ]
Min, Geyong [1 ]
Zomaya, Albert Y. [2 ]
El-Ghazawi, Tarek [3 ]
机构
[1] Univ Exeter, Coll Engn Math & Phys Sci, Dept Comp Sci, Exeter EX4 4QF, Devon, England
[2] Univ Sydney, Sch Comp Sci, Camperdown, NSW 2006, Australia
[3] George Washington Univ, Dept Elect & Comp Engn, Washington, DC 20052 USA
关键词
Cloud computing; workload prediction; resource provisioning; sequential data analysis; deep learning; PERFORMANCE; NETWORK; ENERGY;
D O I
10.1109/TPDS.2019.2953745
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Resource provisioning for cloud computing necessitates the adaptive and accurate prediction of cloud workloads. However, the existing methods cannot effectively predict the high-dimensional and highly-variable cloud workloads. This results in resource wasting and inability to satisfy service level agreements (SLAs). Since recurrent neural network (RNN) is naturally suitable for sequential data analysis, it has been recently used to tackle the problem of workload prediction. However, RNN often performs poorly on learning long-term memory dependencies, and thus cannot make the accurate prediction of workloads. To address these important challenges, we propose a deep Learning based Prediction Algorithm for cloud Workloads (L-PAW). First, a top-sparse auto-encoder (TSA) is designed to effectively extract the essential representations of workloads from the original high-dimensional workload data. Next, we integrate TSA and gated recurrent unit (GRU) block into RNN to achieve the adaptive and accurate prediction for highly-variable workloads. Using real-world workload traces from Google and Alibaba cloud data centers and the DUX-based cluster, extensive experiments are conducted to demonstrate the effectiveness and adaptability of the L-PAW for different types of workloads with various prediction lengths. Moreover, the performance results show that the L-PAW achieves superior prediction accuracy compared to the classic RNN-based and other workload prediction methods for high-dimensional and highly-variable real-world cloud workloads.
引用
收藏
页码:923 / 934
页数:12
相关论文
共 36 条
  • [1] Abadi M, 2016, ACM SIGPLAN NOTICES, V51, P1, DOI [10.1145/2951913.2976746, 10.1145/3022670.2976746]
  • [2] Adaptive Resource Allocation and Provisioning in Multi-Service Cloud Environments
    Alsarhan, Ayoub
    Itradat, Awni
    Al-Dubai, Ahmed Y.
    Zomaya, Albert Y.
    Min, Geyong
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2018, 29 (01) : 31 - 42
  • [3] [Anonymous], 2018, J SUPERCOMPUT, DOI DOI 10.1007/s11227-017-2044-4
  • [4] Bengio Y, 2014, P NIPS WORKSH DEEP L
  • [5] Learning-Based Resource Allocation in Cloud Data Center Using Advantage Actor-Critic
    Chen, Zheyi
    Hu, Jia
    Min, Geyong
    [J]. ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2019,
  • [6] Resource Central: Understanding and Predicting Workloads for Improved Resource Management in Large Cloud Platforms
    Cortez, Eli
    Bonde, Anand
    Muzio, Alexandre
    Russinovich, Mark
    Fontoura, Marcus
    Bianchini, Ricardo
    [J]. PROCEEDINGS OF THE TWENTY-SIXTH ACM SYMPOSIUM ON OPERATING SYSTEMS PRINCIPLES (SOSP '17), 2017, : 153 - 167
  • [7] Paragon: QoS-Aware Scheduling for Heterogeneous Datacenters
    Delimitrou, Christina
    Kozyrakis, Christos
    [J]. ACM SIGPLAN NOTICES, 2013, 48 (04) : 77 - 88
  • [8] Dinda P. A., 1999, Scientific Programming, V7, P211
  • [9] Duggan M, 2017, INT CONF INTERNET, P67, DOI 10.23919/ICITST.2017.8356348
  • [10] Goodfellow I., 2016, DEEP LEARNING