HephaestusForge: Optimal microservice deployment across the Compute Continuum via Reinforcement Learning

被引:0
|
作者
Santos, Jose [1 ]
Zaccarini, Mattia [2 ]
Poltronieri, Filippo [2 ]
Tortonesi, Mauro [2 ]
Stefanelli, Cesare [2 ]
Di Cicco, Nicola [3 ]
De Turck, Filip [1 ]
机构
[1] Univ Ghent, Dept Informat Technol, IDLab, Imec, Technol Pk Zwijnaarde 126, B-9052 Ghent, Belgium
[2] Univ Ferrara, Distributed Syst Res Grp, Ferrara, Italy
[3] Politecn Milan, Dept Elect Informat & Bioengn DEIB, Milan, Italy
来源
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE | 2025年 / 166卷
关键词
Kubernetes; Orchestration; Microservices; Reinforcement Learning; Resource allocation; Compute Continuum; SERVICE FUNCTION CHAIN; CLOUD; ORCHESTRATION;
D O I
10.1016/j.future.2024.107680
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
With the advent of containerization technologies, microservices have revolutionized application deployment by converting old monolithic software into a group of loosely coupled containers, aiming to offer greater flexibility and improve operational efficiency. This transition made applications more complex, consisting of tens to hundreds of microservices. Designing effective orchestration mechanisms remains a crucial challenge, especially for emerging distributed cloud paradigms such as the Compute Continuum (CC). Orchestration across multiple clusters is still not extensively explored in the literature since most works consider single- cluster scenarios. In the CC scenario, the orchestrator must decide the optimal locations for each microservice, deciding whether instances are deployed altogether or placed across different clusters, significantly increasing orchestration complexity. This paper addresses orchestration in a containerized CC environment by studying a Reinforcement Learning (RL) approach for efficient microservice deployment in Kubernetes (K8s) clusters, a widely adopted container orchestration platform. This work demonstrates the effectiveness of RL in achieving near-optimal deployment schemes under dynamic conditions, where network latency and resource capacity fluctuate. We extensively evaluate a multi-objective reward function that aims to minimize overall latency, reduce deployment costs, and promote fair distribution of microservice instances, and we compare it against typical heuristic-based approaches. The results from an implemented OpenAI Gym framework, named as HephaestusForge, show that RL algorithms achieve minimal rejection rates (as low as 0.002%, 90x less than the baseline Karmada scheduler). Cost-aware strategies result in lower deployment costs (2.5 units), and latency- aware functions achieve lower latency (268-290 ms), improving by 1.5x and 1.3x, respectively, over the best-performing baselines. HephaestusForge is available in a public open-source repository, allowing researchers to validate their own placement algorithms. This study also highlights the adaptability of the DeepSets (DS) neural network in optimizing microservice placement across diverse multi-cluster setups without retraining. The DS neural network can handle inputs and outputs as arbitrarily sized sets, enabling the RL algorithm to learn a policy not bound to a fixed number of clusters.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] Optimal online energy management strategy of a fuel cell hybrid bus via reinforcement learning
    Deng, Pengyi
    Wu, Xiaohua
    Yang, Jialuo
    Yang, Gang
    Jiang, Ping
    Yang, Jibin
    Bian, Xiaolei
    ENERGY CONVERSION AND MANAGEMENT, 2024, 300
  • [32] Optimal Adaptive Prediction Intervals for Electricity Load Forecasting in Distribution Systems via Reinforcement Learning
    Zhang, Yufan
    Wen, Honglin
    Wu, Qiuwei
    Ai, Qian
    IEEE TRANSACTIONS ON SMART GRID, 2023, 14 (04) : 3259 - 3270
  • [33] Coordinated Ship Welding with Optimal Lazy Robot Ratio and Energy Consumption via Reinforcement Learning
    Yu, Rui
    Chen, Yang-Yang
    JOURNAL OF MARINE SCIENCE AND ENGINEERING, 2024, 12 (10)
  • [34] Adaptive Neural Network Optimal Backstepping Control of Strict Feedback Nonlinear Systems via Reinforcement Learning
    Zhong, Mei
    Cao, Jinde
    Liu, Heng
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2025, 9 (01): : 832 - 847
  • [35] Reinforcement Learning-Based Cooperative Optimal Output Regulation via Distributed Adaptive Internal Model
    Gao, Weinan
    Mynuddin, Mohammed
    Wunsch, Donald C.
    Jiang, Zhong-Ping
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (10) : 5229 - 5240
  • [36] Optimal Robust Control of Nonlinear Uncertain System via Off-Policy Integral Reinforcement Learning
    Wang, Xiaoyang
    Ye, Xiufen
    PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 1928 - 1933
  • [37] Optimal bipartite consensus control for heterogeneous unknown multi-agent systems via reinforcement learning
    Meng, Hao
    Pang, Denghao
    Cao, Jinde
    Guo, Yechen
    Niazi, Azmat Ullah Khan
    APPLIED MATHEMATICS AND COMPUTATION, 2024, 476
  • [38] HOODIE: Hybrid Computation Offloading via Distributed Deep Reinforcement Learning in Delay-Aware Cloud-Edge Continuum
    Giannopoulos, Anastasios E.
    Paralikas, Ilias
    Spantideas, Sotirios T.
    Trakadas, Panagiotis
    IEEE OPEN JOURNAL OF THE COMMUNICATIONS SOCIETY, 2024, 5 : 7818 - 7841
  • [39] Delay-Optimal Scheduling for Heavy-Tailed and Light-Tailed Flows via Reinforcement Learning
    Guo, Mian
    Guan, Quansheng
    Chen, Weiqi
    Ji, Fei
    Peng, Zhiping
    PROCEEDINGS OF 2018 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATION SYSTEMS (ICCS 2018), 2018, : 292 - 296
  • [40] Data-driven optimal tracking control for SMA actuated systems with prescribed performance via reinforcement learning
    Liu, Hongshuai
    Cheng, Qiang
    Xiao, Jichun
    Hao, Lina
    MECHANICAL SYSTEMS AND SIGNAL PROCESSING, 2022, 177