Analysis of Resource Management Methods Based on Reinforcement Learning

被引:0
作者
Xing, Mingzhe [1 ]
Wang, Ziyun [2 ]
Xiao, Zhen [1 ]
机构
[1] Peking Univ, Dept Comp Sci, Beijing, Peoples R China
[2] Natl Univ Singapore, Fac Sci, Singapore, Singapore
来源
2021 INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE BIG DATA AND INTELLIGENT SYSTEMS (HPBD&IS) | 2021年
基金
中国国家自然科学基金;
关键词
Deep Learning; Reinforcement Learning; Resource Management; Graph Neural Networks; GO;
D O I
10.1109/HPBDIS53214.2021.9658350
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As the scale of service-based applications rapidly grows over recent years, tremendous amount of user data is being generated on a daily basis, which needs to be processed by computing jobs. Distributed computing frameworks are extensively applied to efficiently process large-scale data using finite resources, which has consequently placed resource management at the center of attention for many researchers. Traditional heuristic-based resource management algorithms are widely used in the industry, while often require experts with rich experience to design and tune rules, which is usually a timeconsuming process and difficult to be generalized to computing jobs with distinct natures and scales. With the immense successes of reinforcement learning (RL) in the fields of games, autodriving, and robotics, researchers begin to model and learn the task of resource management through the perspectives of RL, which has been proven to outperform conventional methods by experimental results. In this paper, we aim to summarize the relevant background, introduce both the heuristic-based and RLbased algorithms and propose a few areas of improvement for future work to come.
引用
收藏
页码:27 / 31
页数:5
相关论文
共 40 条
  • [1] A survey of current challenges in partitioning and processing of graph-structured data in parallel and distributed systems
    Adoni, Hamilton Wilfried Yves
    Nahhal, Tarik
    Krichen, Moez
    Aghezzaf, Brahim
    Elbyed, Abdeltif
    [J]. DISTRIBUTED AND PARALLEL DATABASES, 2020, 38 (02) : 495 - 530
  • [2] [Anonymous], 2011, PROC USENIX C NETWOR
  • [3] Long short-term memory
    Hochreiter, S
    Schmidhuber, J
    [J]. NEURAL COMPUTATION, 1997, 9 (08) : 1735 - 1780
  • [4] Busoniu L, 2010, STUD COMPUT INTELL, V310, P183
  • [5] Du BQ, 2019, AAAI CONF ARTIF INTE, P7570
  • [6] Fahs AJ, 2020, I S MOD ANAL SIM COM, P168
  • [7] Reinforcement learning-based application Autoscaling in the Cloud: A survey
    Gari, Yisel
    Monge, David A.
    Pacini, Elina
    Mateos, Cristian
    Garcia Garino, Carlos
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2021, 102
  • [8] Grandl R, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P81
  • [9] Henaff M, 2015, ARXIV150605163
  • [10] Hua Wei, 2020, ACM SIGKDD Explorations Newsletter, V22, P12, DOI 10.1145/3447556.3447565