DeepMRA: An Efficient Microservices Resource Allocation Framework with Deep Reinforcement Learning in the Cloud

被引:0
作者
Si, Qi [1 ]
Shi, Jilin [1 ]
Li, Weiyi [1 ]
Lu, Xuesong [1 ]
Pu, Peng [1 ]
机构
[1] East China Normal Univ, Sch Data Sci & Engn, Shanghai 200062, Peoples R China
来源
ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT II, ICIC 2024 | 2024年 / 14863卷
关键词
Resource Allocation; Deep Reinforcement Learning; Cloud Computing; Microservice;
D O I
10.1007/978-981-97-5581-3_37
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The rapid growth of cloud computing has precipitated a paradigm shift in application service deployment, transitioning predominantly from monolithic to microservices architectures. This shift to microservices brings new, complex challenges in managing cloud resources. Traditional cloud resource allocation methods struggle with microservices' unique challenges, like complex inter-service dependencies and the need to balance Quality of Service (QoS) with cost efficiency. Recognizing these challenges in cloud resource management, our study proposes an innovative approach to dynamically allocate resources for cloud microservices with Deep Reinforcement Learning (DRL). Specifically, we introduce DeepMRA, an efficient microservices resource allocation framework with Deep Reinforcement Learning in the Cloud, with multiple agents navigating the complexities arising from varying workloads. We propose a performance predictor to forecast application performance, guiding the training of agents in DRL. Due to the shortcomings of traditional performance data collection methods in the context of microservices, we developed the Parallel and Asynchronous Uncertainty-Directed Sampling (PAUDS) algorithm. This algorithm is specifically designed to optimize data collection processes, ensuring a robust dataset for building a reliable performance predictor. Extensive experiments conducted with microservice-based applications indicate that the proposed method reduces resource consumption while upholding QoS requirements under varying workloads.
引用
收藏
页码:455 / 466
页数:12
相关论文
共 23 条
  • [1] autodesk, About us
  • [2] Adaptive and Efficient Resource Allocation in Cloud Datacenters Using Actor-Critic Deep Reinforcement Learning
    Chen, Zheyi
    Hu, Jia
    Min, Geyong
    Luo, Chunbo
    El-Ghazawi, Tarek
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2022, 33 (08) : 1911 - 1923
  • [3] Paragon: QoS-Aware Scheduling for Heterogeneous Datacenters
    Delimitrou, Christina
    Kozyrakis, Christos
    [J]. ACM SIGPLAN NOTICES, 2013, 48 (04) : 77 - 88
  • [4] docs.aws.amazon, EC2 Auto Scaling
  • [5] Dong H, 2021, PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, P3627
  • [6] SPEEDUP VERSUS EFFICIENCY IN PARALLEL SYSTEMS
    EAGER, DL
    ZAHORJAN, J
    LAZOWSKA, ED
    [J]. IEEE TRANSACTIONS ON COMPUTERS, 1989, 38 (03) : 408 - 423
  • [7] Fan YP, 2021, Arxiv, DOI [arXiv:2102.06243, DOI 10.48550/ARXIV.2102.06243]
  • [8] An Open-Source Benchmark Suite for Microservices and Their Hardware-Software Implications for Cloud & Edge Systems
    Gan, Yu
    Zhang, Yanqi
    Cheng, Dailun
    Shetty, Ankitha
    Rathi, Priyal
    Katarki, Nayan
    Bruno, Ariana
    Hu, Justin
    Ritchken, Brian
    Jackson, Brendon
    Hu, Kelvin
    Pancholi, Meghna
    He, Yuan
    Clancy, Brett
    Colen, Chris
    Wen, Fukang
    Leung, Catherine
    Wang, Siyuan
    Zaruvinsky, Leon
    Espinosa, Mateo
    Lin, Rick
    Liu, Zhongling
    Padilla, Jake
    Delimitrou, Christina
    [J]. TWENTY-FOURTH INTERNATIONAL CONFERENCE ON ARCHITECTURAL SUPPORT FOR PROGRAMMING LANGUAGES AND OPERATING SYSTEMS (ASPLOS XXIV), 2019, : 3 - 18
  • [9] Haarnoja T, 2018, Arxiv, DOI [arXiv:1801.01290, 10.48550/arXiv.1801.01290, DOI 10.48550/ARXIV.1801.01290]
  • [10] Performance and Cost-Efficient Spark Job Scheduling Based on Deep Reinforcement Learning in Cloud Computing Environments
    Islam, Muhammed Tawfiqul
    Karunasekera, Shanika
    Buyya, Rajkumar
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2022, 33 (07) : 1695 - 1710