HephaestusForge: Optimal microservice deployment across the Compute Continuum via Reinforcement Learning

被引:0
|
作者
Santos, Jose [1 ]
Zaccarini, Mattia [2 ]
Poltronieri, Filippo [2 ]
Tortonesi, Mauro [2 ]
Stefanelli, Cesare [2 ]
Di Cicco, Nicola [3 ]
De Turck, Filip [1 ]
机构
[1] Univ Ghent, Dept Informat Technol, IDLab, Imec, Technol Pk Zwijnaarde 126, B-9052 Ghent, Belgium
[2] Univ Ferrara, Distributed Syst Res Grp, Ferrara, Italy
[3] Politecn Milan, Dept Elect Informat & Bioengn DEIB, Milan, Italy
来源
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE | 2025年 / 166卷
关键词
Kubernetes; Orchestration; Microservices; Reinforcement Learning; Resource allocation; Compute Continuum; SERVICE FUNCTION CHAIN; CLOUD; ORCHESTRATION;
D O I
10.1016/j.future.2024.107680
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
With the advent of containerization technologies, microservices have revolutionized application deployment by converting old monolithic software into a group of loosely coupled containers, aiming to offer greater flexibility and improve operational efficiency. This transition made applications more complex, consisting of tens to hundreds of microservices. Designing effective orchestration mechanisms remains a crucial challenge, especially for emerging distributed cloud paradigms such as the Compute Continuum (CC). Orchestration across multiple clusters is still not extensively explored in the literature since most works consider single- cluster scenarios. In the CC scenario, the orchestrator must decide the optimal locations for each microservice, deciding whether instances are deployed altogether or placed across different clusters, significantly increasing orchestration complexity. This paper addresses orchestration in a containerized CC environment by studying a Reinforcement Learning (RL) approach for efficient microservice deployment in Kubernetes (K8s) clusters, a widely adopted container orchestration platform. This work demonstrates the effectiveness of RL in achieving near-optimal deployment schemes under dynamic conditions, where network latency and resource capacity fluctuate. We extensively evaluate a multi-objective reward function that aims to minimize overall latency, reduce deployment costs, and promote fair distribution of microservice instances, and we compare it against typical heuristic-based approaches. The results from an implemented OpenAI Gym framework, named as HephaestusForge, show that RL algorithms achieve minimal rejection rates (as low as 0.002%, 90x less than the baseline Karmada scheduler). Cost-aware strategies result in lower deployment costs (2.5 units), and latency- aware functions achieve lower latency (268-290 ms), improving by 1.5x and 1.3x, respectively, over the best-performing baselines. HephaestusForge is available in a public open-source repository, allowing researchers to validate their own placement algorithms. This study also highlights the adaptability of the DeepSets (DS) neural network in optimizing microservice placement across diverse multi-cluster setups without retraining. The DS neural network can handle inputs and outputs as arbitrarily sized sets, enabling the RL algorithm to learn a policy not bound to a fixed number of clusters.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] EV charging station deployment on coupled transportation and power distribution networks via reinforcement learning
    Zhao, Zhonghao
    Lee, Carman K. M.
    Huo, Jiage
    ENERGY, 2023, 267
  • [22] An efficient integration of the genetic algorithm and the reinforcement learning for optimal deployment of the wireless charging electric tram system
    Ko, Young Dae
    COMPUTERS & INDUSTRIAL ENGINEERING, 2019, 128 : 851 - 860
  • [23] Optimal Policy of Multiplayer Poker via Actor-Critic Reinforcement Learning
    Shi, Daming
    Guo, Xudong
    Liu, Yi
    Fan, Wenhui
    ENTROPY, 2022, 24 (06)
  • [24] Optimal redundant transmission scheduling for remote state estimation via reinforcement learning approach
    Jia, Yijin
    Yang, Lixin
    Zhao, Yao
    Li, Jun-Yi
    Lv, Weijun
    NEUROCOMPUTING, 2024, 576
  • [25] Using process data to generate an optimal control policy via apprenticeship and reinforcement learning
    Mowbray, Max
    Smith, Robin
    Del Rio-Chanona, Ehecatl A.
    Zhang, Dongda
    AICHE JOURNAL, 2021, 67 (09)
  • [26] Adaptive Optimal Control via Reinforcement Learning for Omni-Directional Wheeled Robots
    Sheikhlar, Arash
    Fakharian, Ahmad
    2016 4TH INTERNATIONAL CONFERENCE ON CONTROL, INSTRUMENTATION, AND AUTOMATION (ICCIA), 2016, : 208 - 213
  • [27] Learning state-action correspondence across reinforcement learning control tasks via partially paired trajectories
    Garcia, Javier
    Rano, Inaki
    Bures, J. Miguel
    Fdez-Vidal, Xose R.
    Iglesias, Roberto
    APPLIED INTELLIGENCE, 2025, 55 (03)
  • [28] End-to-End Active Object Tracking and Its Real-World Deployment via Reinforcement Learning
    Luo, Wenhan
    Sun, Peng
    Zhong, Fangwei
    Liu, Wei
    Zhang, Tong
    Wang, Yizhou
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (06) : 1317 - 1332
  • [29] POMDP inference and robust solution via deep reinforcement learning: an application to railway optimal maintenance
    Arcieri, Giacomo
    Hoelzl, Cyprien
    Schwery, Oliver
    Straub, Daniel
    Papakonstantinou, Konstantinos G.
    Chatzi, Eleni
    MACHINE LEARNING, 2024, 113 (10) : 7967 - 7995
  • [30] Robust Optimal Formation Control of Heterogeneous Multi-Agent System via Reinforcement Learning
    Lin, Wei
    Zhao, Wanbing
    Liu, Hao
    IEEE ACCESS, 2020, 8 (08): : 218424 - 218432