A Deep Recurrent-Reinforcement Learning Method for Intelligent AutoScaling of Serverless Functions

被引:4
|
作者
Agarwal, Siddharth [1 ]
Rodriguez, Maria A. [1 ]
Buyya, Rajkumar [1 ]
机构
[1] Univ Melbourne, Sch Comp & Informat Syst, Cloud Comp & Distributed Syst CLOUDS Lab, Melbourne, Vic 3010, Australia
关键词
Serverless computing; function-as-a-service; AutoScaling; reinforcement learning; constraint-awareness;
D O I
10.1109/TSC.2024.3387661
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Function-as-a-Service (FaaS) introduces a lightweight, function-based cloud execution model that finds its relevance in a range of applications like IoT-edge data processing and anomaly detection. While cloud service providers (CSPs) offer a near-infinite function elasticity, these applications often experience fluctuating workloads and stricter performance constraints. A typical CSP strategy is to empirically determine and adjust desired function instances or resources, known as autoscaling, based on monitoring-based thresholds such as CPU or memory, to cope with demand and performance. However, threshold configuration either requires expert knowledge, historical data or a complete view of the environment, making autoscaling a performance bottleneck that lacks an adaptable solution. Reinforcement learning (RL) algorithms are proven to be beneficial in analysing complex cloud environments and result in an adaptable policy that maximizes the expected objectives. Most realistic cloud environments usually involve operational interference and have limited visibility, making them partially observable. A general solution to tackle observability in highly dynamic settings is to integrate Recurrent units with model-free RL algorithms and model a decision process as a Partially Observable Markov Decision Process (POMDP). Therefore, in this article, we investigate model-free Recurrent RL agents for function autoscaling and compare them against the model-free Proximal Policy Optimisation (PPO) algorithm. We explore the integration of a Long-Short Term Memory (LSTM) network with the state-of-the-art PPO algorithm to find that under our experimental and evaluation settings, recurrent policies were able to capture the environment parameters and show promising results for function autoscaling. We further compare a PPO-based autoscaling agent with commercially used threshold-based function autoscaling and posit that a LSTM-based autoscaling agent is able to improve throughput by 18%, function execution by 13% and account for 8.4% more function instances.
引用
收藏
页码:1899 / 1910
页数:12
相关论文
共 50 条
  • [1] Reinforcement learning-assisted autoscaling mechanisms for serverless computing platforms
    Zafeiropoulos, Anastasios
    Fotopoulou, Eleni
    Filinis, Nikos
    Papavassiliou, Symeon
    SIMULATION MODELLING PRACTICE AND THEORY, 2022, 116
  • [2] Intelligent microservices autoscaling module using reinforcement learning
    Abeer Abdel Khaleq
    Ilkyeun Ra
    Cluster Computing, 2023, 26 : 2789 - 2800
  • [3] Intelligent microservices autoscaling module using reinforcement learning
    Khaleq, Abeer Abdel
    Ra, Ilkyeun
    CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2023, 26 (05): : 2789 - 2800
  • [4] Freyr+: Harvesting Idle Resources in Serverless Computing via Deep Reinforcement Learning
    Yu, Hanfei
    Wang, Hao
    Li, Jian
    Yuan, Xu
    Park, Seung-Jong
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2024, 35 (11) : 2254 - 2269
  • [5] A scalable modified deep reinforcement learning algorithm for serverless IoT microservice composition infrastructure in fog layer
    Khansari, Mina Emami
    Sharifian, Saeed
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 153 : 206 - 221
  • [6] Priority-Aware Deployment of Autoscaling Service Function Chains Based on Deep Reinforcement Learning
    Yu, Xue
    Wang, Ran
    Hao, Jie
    Wu, Qiang
    Yi, Changyan
    Wang, Ping
    Niyato, Dusit
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2024, 10 (03) : 1050 - 1062
  • [7] Deep Reinforcement Learning with Successive Over-Relaxation and its Application in Autoscaling Cloud Resources
    John, Indu
    Bhatnagar, Shalabh
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [8] Intelligent land vehicle model transfer trajectory planning method of deep reinforcement learning
    Yu L.-L.
    Shao X.-Y.
    Long Z.-W.
    Wei Y.-D.
    Zhou K.-J.
    Kongzhi Lilun Yu Yingyong/Control Theory and Applications, 2019, 36 (09): : 1409 - 1422
  • [9] Performance optimization of serverless edge computing function offloading based on deep reinforcement learning
    Yao, Xuyi
    Chen, Ningjiang
    Yuan, Xuemei
    Ou, Pingjie
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2023, 139 : 74 - 86
  • [10] Multi-Objective Deep Reinforcement Learning for Function Offloading in Serverless Edge Computing
    Yang, Yaning
    Du, Xiao
    Ye, Yutong
    Ding, Jiepin
    Wang, Ting
    Chen, Mingsong
    Li, Keqin
    IEEE TRANSACTIONS ON SERVICES COMPUTING, 2025, 18 (01) : 288 - 301