HephaestusForge: Optimal microservice deployment across the Compute Continuum via Reinforcement Learning

被引:0
|
作者
Santos, Jose [1 ]
Zaccarini, Mattia [2 ]
Poltronieri, Filippo [2 ]
Tortonesi, Mauro [2 ]
Stefanelli, Cesare [2 ]
Di Cicco, Nicola [3 ]
De Turck, Filip [1 ]
机构
[1] Univ Ghent, Dept Informat Technol, IDLab, Imec, Technol Pk Zwijnaarde 126, B-9052 Ghent, Belgium
[2] Univ Ferrara, Distributed Syst Res Grp, Ferrara, Italy
[3] Politecn Milan, Dept Elect Informat & Bioengn DEIB, Milan, Italy
来源
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE | 2025年 / 166卷
关键词
Kubernetes; Orchestration; Microservices; Reinforcement Learning; Resource allocation; Compute Continuum; SERVICE FUNCTION CHAIN; CLOUD; ORCHESTRATION;
D O I
10.1016/j.future.2024.107680
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
With the advent of containerization technologies, microservices have revolutionized application deployment by converting old monolithic software into a group of loosely coupled containers, aiming to offer greater flexibility and improve operational efficiency. This transition made applications more complex, consisting of tens to hundreds of microservices. Designing effective orchestration mechanisms remains a crucial challenge, especially for emerging distributed cloud paradigms such as the Compute Continuum (CC). Orchestration across multiple clusters is still not extensively explored in the literature since most works consider single- cluster scenarios. In the CC scenario, the orchestrator must decide the optimal locations for each microservice, deciding whether instances are deployed altogether or placed across different clusters, significantly increasing orchestration complexity. This paper addresses orchestration in a containerized CC environment by studying a Reinforcement Learning (RL) approach for efficient microservice deployment in Kubernetes (K8s) clusters, a widely adopted container orchestration platform. This work demonstrates the effectiveness of RL in achieving near-optimal deployment schemes under dynamic conditions, where network latency and resource capacity fluctuate. We extensively evaluate a multi-objective reward function that aims to minimize overall latency, reduce deployment costs, and promote fair distribution of microservice instances, and we compare it against typical heuristic-based approaches. The results from an implemented OpenAI Gym framework, named as HephaestusForge, show that RL algorithms achieve minimal rejection rates (as low as 0.002%, 90x less than the baseline Karmada scheduler). Cost-aware strategies result in lower deployment costs (2.5 units), and latency- aware functions achieve lower latency (268-290 ms), improving by 1.5x and 1.3x, respectively, over the best-performing baselines. HephaestusForge is available in a public open-source repository, allowing researchers to validate their own placement algorithms. This study also highlights the adaptability of the DeepSets (DS) neural network in optimizing microservice placement across diverse multi-cluster setups without retraining. The DS neural network can handle inputs and outputs as arbitrarily sized sets, enabling the RL algorithm to learn a policy not bound to a fixed number of clusters.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Optimal Resource Allocation for Reconfigurable Intelligent Surface Assisted Dynamic Wireless Network via Online Reinforcement Learning
    Zhang, Yuzhu
    Xu, Hao
    2022 IEEE INTERNATIONAL CONFERENCE ON SENSING, COMMUNICATION, AND NETWORKING (SECON WORKSHOPS), 2022, : 13 - 18
  • [42] Optimal Containment Control for Unknown Active Heterogeneous MASs via Model-Free Recursive Reinforcement Learning
    Xia, Lina
    Li, Qing
    Song, Ruizhuo
    Yang, Gaofu
    IEEE ACCESS, 2025, 13 : 7603 - 7613
  • [43] Data-driven Optimal Control Strategy for Virtual Synchronous Generator via Deep Reinforcement Learning Approach
    Li, Yushuai
    Gao, Wei
    Yan, Weihang
    Huang, Shuo
    Wang, Rui
    Gevorgian, Vahan
    Gao, David Wenzhong
    JOURNAL OF MODERN POWER SYSTEMS AND CLEAN ENERGY, 2021, 9 (04) : 919 - 929
  • [44] Attack-Resilient Optimal PMU Placement via Reinforcement Learning Guided Tree Search in Smart Grids
    Zhang, Meng
    Wu, Zhuorui
    Yan, Jun
    Lu, Rongxing
    Guan, Xiaohong
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2022, 17 : 1919 - 1929
  • [45] Pareto-optimal synchronization control of nonlinear multi-agent systems via integral reinforcement learning
    Guo, Yaning
    Sun, Qi
    Pan, Quan
    Wang, Yintao
    NONLINEAR DYNAMICS, 2025, 113 (06) : 5339 - 5357
  • [46] A Data-driven Method for Fast AC Optimal Power Flow Solutions via Deep Reinforcement Learning
    Zhou, Yuhao
    Zhang, Bei
    Xu, Chunlei
    Lan, Tu
    Diao, Ruisheng
    Shi, Di
    Wang, Zhiwei
    Lee, Wei-Jen
    JOURNAL OF MODERN POWER SYSTEMS AND CLEAN ENERGY, 2020, 8 (06) : 1128 - 1139
  • [47] OPTIMAL SYNCHRONIZATION CONTROL OF MULTIPLE EULER-LAGRANGE SYSTEMS VIA EVENT-TRIGGERED REINFORCEMENT LEARNING
    Xu, Yuan
    Jin, Xin
    Wang, Saiwei
    Tang, Yang
    DISCRETE AND CONTINUOUS DYNAMICAL SYSTEMS-SERIES S, 2021, 14 (04): : 1495 - 1518
  • [48] Mixed-Integer Optimal Control via Reinforcement Learning: A Case Study on Hybrid Electric Vehicle Energy Management
    Xu, Jinming
    Azad, Nasser Lashgarian
    Lin, Yuan
    OPTIMAL CONTROL APPLICATIONS & METHODS, 2025, 46 (01) : 307 - 319
  • [49] Adaptive Neural Optimal Backstepping Control of Uncertain Fractional-Order Chaotic Circuit Systems via Reinforcement Learning
    Zhong, Mei
    Huang, Chengdai
    Cao, Jinde
    Liu, Heng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2024, 71 (10) : 4707 - 4720
  • [50] Adaptive optimal output tracking of continuous-time systems via output-feedback-based reinforcement learning
    Chen, Ci
    Xie, Lihua
    Xie, Kan
    Lewis, Frank L.
    Xie, Shengli
    AUTOMATICA, 2022, 146