Online Container Scheduling With Fast Function Startup and Low Memory Cost in Edge Computing

被引:0
作者
Li, Zhenzheng [1 ,2 ]
Lou, Jiong [3 ]
Wu, Jianfei [1 ,2 ]
Guo, Jianxiong [2 ,4 ]
Tang, Zhiqing [2 ]
Shen, Ping [2 ]
Jia, Weijia [2 ,4 ]
Zhao, Wei [5 ]
机构
[1] Beijing Normal Univ, Sch Artificial Intelligence, Beijing 100875, Peoples R China
[2] Beijing Normal Univ, Inst Artificial Intelligence & Future Networks, Zhuhai 519087, Peoples R China
[3] Shanghai Jiao Tong Univ, Dept Comp Sci & Engn, Shanghai 200240, Peoples R China
[4] BNU HKBU United Int Coll, Guangdong Key Lab AI & Multimodal Data Proc, Zhuhai 519087, Peoples R China
[5] Shenzhen Univ Adv Technol, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Containers; Delays; Costs; Memory management; Serverless computing; Processor scheduling; Optimization; zygote container; scheduling; online optimization; ALGORITHMS; RENT; BUY;
D O I
10.1109/TC.2024.3441836
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Extending serverless computing to the edge has emerged as a promising approach to support service, but startup containerized serverless functions lead to the cold-start delay. Recent research has introduced container caching methods to alleviate the cold-start delay, including cache as the entire container or the Zygote container. However, container caching incurs memory costs. The system must ensure fast function startup and low memory cost of edge servers, which has been overlooked in the literature. This paper aims to jointly optimize startup delay and memory cost. We formulate an online joint optimization problem that encompasses container scheduling decisions, including invocation distribution, container startup, and container caching. To solve the problem, we propose an online algorithm with a competitive ratio and low computational complexity. The proposed algorithm decomposes the problem into two subproblems and solves them sequentially. Each container is assigned a randomized strategy, and these container-level decisions are merged to constitute overall container caching decisions. Furthermore, a greedy-based subroutine is designed to solve the subproblem associated with invocation distribution and container startup decisions. Experiments on the real-world dataset indicate that the algorithm can reduce average startup delay by up to 23% and lower memory costs by up to 15%.
引用
收藏
页码:2747 / 2760
页数:14
相关论文
共 41 条
  • [1] Akkus IE, 2018, PROCEEDINGS OF THE 2018 USENIX ANNUAL TECHNICAL CONFERENCE, P923
  • [2] [Anonymous], MICROSOFT AZURE FUNC
  • [3] [Anonymous], AWS LAMBDA
  • [4] Apache openwhisk, ABOUT US
  • [5] Arutyunyan M, 2023, PROCEEDINGS OF THE 2023 USENIX ANNUAL TECHNICAL CONFERENCE, P329
  • [6] Optimal power-down strategies
    Augustine, J
    Irani, S
    Swamy, C
    [J]. 45TH ANNUAL IEEE SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE, PROCEEDINGS, 2004, : 530 - 539
  • [7] On randomization in on-line computation
    Borodin, A
    El-Yaniv, R
    [J]. INFORMATION AND COMPUTATION, 1999, 150 (02) : 244 - 267
  • [8] Boyd SP., 2004, CONVEX OPTIMIZATION, DOI [DOI 10.1017/CBO9780511804441, 10.1017/CBO9780511804441]
  • [9] Xanadu: Mitigating cascading cold starts in serverless function chain deployments
    Daw, Nilanjan
    Bellur, Umesh
    Kulkarni, Purushottam
    [J]. PROCEEDINGS OF THE 2020 21ST INTERNATIONAL MIDDLEWARE CONFERENCE (MIDDLEWARE '20), 2020, : 356 - 370
  • [10] Dong CS, 2014, DES AUT CON