A Deep Recurrent-Reinforcement Learning Method for Intelligent AutoScaling of Serverless Functions

被引:4
|
作者
Agarwal, Siddharth [1 ]
Rodriguez, Maria A. [1 ]
Buyya, Rajkumar [1 ]
机构
[1] Univ Melbourne, Sch Comp & Informat Syst, Cloud Comp & Distributed Syst CLOUDS Lab, Melbourne, Vic 3010, Australia
关键词
Serverless computing; function-as-a-service; AutoScaling; reinforcement learning; constraint-awareness;
D O I
10.1109/TSC.2024.3387661
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Function-as-a-Service (FaaS) introduces a lightweight, function-based cloud execution model that finds its relevance in a range of applications like IoT-edge data processing and anomaly detection. While cloud service providers (CSPs) offer a near-infinite function elasticity, these applications often experience fluctuating workloads and stricter performance constraints. A typical CSP strategy is to empirically determine and adjust desired function instances or resources, known as autoscaling, based on monitoring-based thresholds such as CPU or memory, to cope with demand and performance. However, threshold configuration either requires expert knowledge, historical data or a complete view of the environment, making autoscaling a performance bottleneck that lacks an adaptable solution. Reinforcement learning (RL) algorithms are proven to be beneficial in analysing complex cloud environments and result in an adaptable policy that maximizes the expected objectives. Most realistic cloud environments usually involve operational interference and have limited visibility, making them partially observable. A general solution to tackle observability in highly dynamic settings is to integrate Recurrent units with model-free RL algorithms and model a decision process as a Partially Observable Markov Decision Process (POMDP). Therefore, in this article, we investigate model-free Recurrent RL agents for function autoscaling and compare them against the model-free Proximal Policy Optimisation (PPO) algorithm. We explore the integration of a Long-Short Term Memory (LSTM) network with the state-of-the-art PPO algorithm to find that under our experimental and evaluation settings, recurrent policies were able to capture the environment parameters and show promising results for function autoscaling. We further compare a PPO-based autoscaling agent with commercially used threshold-based function autoscaling and posit that a LSTM-based autoscaling agent is able to improve throughput by 18%, function execution by 13% and account for 8.4% more function instances.
引用
收藏
页码:1899 / 1910
页数:12
相关论文
共 50 条
  • [41] Mixed Deep Reinforcement Learning-behavior Tree for Intelligent Agents Design
    Li, Lei
    Wang, Lei
    Li, Yuanzhi
    Sheng, Jie
    ICAART: PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 1, 2021, : 113 - 124
  • [42] An Intelligent Framework for English Teaching through Deep Learning and Reinforcement Learning with Interactive Mobile Technology
    Hu J.
    Jin G.
    Int. J. Interact. Mob. Technol., 2024, 9 (74-87): : 74 - 87
  • [43] A Reliable Energy Trading Strategy in Intelligent Microgrids Using Deep Reinforcement Learning
    Cao, Man
    Yin, Zhiyong
    Wang, Yajun
    Yu, Le
    Shi, Peiran
    Cai, Zhi
    COMPUTERS & ELECTRICAL ENGINEERING, 2023, 110
  • [44] When mmWave High-Speed Railway Networks Meet Reconfigurable Intelligent Surface: A Deep Reinforcement Learning Method
    Xu, Jianpeng
    Ai, Bo
    IEEE WIRELESS COMMUNICATIONS LETTERS, 2022, 11 (03) : 533 - 537
  • [45] On Deep Recurrent Reinforcement Learning for Active Visual Tracking of Space Noncooperative Objects
    Zhou, Dong
    Sun, Guanghui
    Zhang, Zhao
    Wu, Ligang
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (08) : 4418 - 4425
  • [46] Integrated Traffic Control for Freeway Recurrent Bottleneck Based on Deep Reinforcement Learning
    Wang, Chong
    Xu, Yang
    Zhang, Jian
    Ran, Bin
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (09) : 15522 - 15535
  • [47] Continuous control with Stacked Deep Dynamic Recurrent Reinforcement Learning for portfolio optimization
    Aboussalah, Amine Mohamed
    Lee, Chi-Guhn
    EXPERT SYSTEMS WITH APPLICATIONS, 2020, 140
  • [48] A new hybrid method of recurrent reinforcement learning and BiLSTM for algorithmic trading
    Huang, Yuling
    Song, Yunlin
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 45 (02) : 1939 - 1951
  • [49] Darly: Deep Reinforcement Learning for QoS-aware scheduling under resource heterogeneity Optimizing serverless video analytics
    Giagkos, Dimitrios
    Tzenetopoulos, Achilleas
    Masouros, Dimosthenis
    Soudris, Dimitrios
    Xydis, Sotirios
    2023 IEEE 16TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING, CLOUD, 2023, : 577 - 579
  • [50] Application of A Deep Reinforcement Learning Method in Financial Market Trading
    Ma, Lixin
    Liu, Yang
    2019 11TH INTERNATIONAL CONFERENCE ON MEASURING TECHNOLOGY AND MECHATRONICS AUTOMATION (ICMTMA 2019), 2019, : 421 - 425