Efficient Online DNN Inference with Continuous Learning in Edge Computing

被引:0
|
作者
Zeng, Yifan [1 ]
Zhou, Ruiting [1 ,2 ]
Jia, Lei [3 ]
Han, Ziyi [1 ]
Yu, Jieling [1 ]
Ma, Yue [2 ]
机构
[1] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan, Peoples R China
[2] Southeast Univ, Sch Comp Sci & Engn, Nanjing, Peoples R China
[3] Univ Oregon, Dept Comp Sci, Eugene, OR 97403 USA
基金
美国国家科学基金会; 中国国家自然科学基金;
关键词
MANAGEMENT;
D O I
10.1109/IWQoS61813.2024.10682835
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Compressed edge DNN models usually experience decreasing model accuracy when performing inference due to data drift. To maintain the inference accuracy, retraining models with continuous learning is usually employed in the edge. However, online edge DNN inference with continuous learning faces new challenges. First, introducing retraining jobs leads to resource competition with the existing edge inference tasks, which will affect the inference latency. Second, retraining jobs and inference tasks exhibit significant differences in workload and latency requirements. These two jobs cannot adopt the same scheduling policy. To overcome the challenges, we propose an Online scheduling algorithm for INference with Continuous learning (OINC). OINC minimizes the weighted sum of the latency of inference tasks and the completion time of retraining jobs with limited edge resources, while ensuring the satisfaction of the inference task's service level objective (SLO) and meeting the deadlines of retraining jobs. OINC first reserves a portion of resources to complete all current inference tasks and allocates the remaining resources to retraining jobs. Subsequently, based on the reserved resource ratio, OINC invokes two sub-algorithms to select edges and allocate resources for each inference task and retraining job respectively. Compared with six state-of-the-art algorithms, OINC can reduce the weighted sum by up to 23.7%, and increase the success rate by up to 35.6%.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Online Optimization of DNN Inference Network Utility in Collaborative Edge Computing
    Li, Rui
    Ouyang, Tao
    Zeng, Liekang
    Liao, Guocheng
    Zhou, Zhi
    Chen, Xu
    IEEE-ACM TRANSACTIONS ON NETWORKING, 2024, 32 (05) : 4414 - 4426
  • [2] DNN Placement and Inference in Edge Computing
    Bensalem, Mounir
    Dizdarevic, Jasenka
    Jukan, Admela
    2020 43RD INTERNATIONAL CONVENTION ON INFORMATION, COMMUNICATION AND ELECTRONIC TECHNOLOGY (MIPRO 2020), 2020, : 479 - 484
  • [3] Resource-Efficient DNN Inference With Early Exiting in Serverless Edge Computing
    Guo, Xiaolin
    Dong, Fang
    Shen, Dian
    Huang, Zhaowu
    Zhang, Jinghui
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2025, 24 (05) : 3650 - 3666
  • [4] Elastic DNN Inference With Unpredictable Exit in Edge Computing
    Huang, Jiaming
    Gao, Yi
    Dong, Wei
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (12) : 14005 - 14016
  • [5] Elastic DNN Inference with Unpredictable Exit in Edge Computing
    Huang, Jiaming
    Gao, Yi
    Dong, Wei
    2023 IEEE 43RD INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS, ICDCS, 2023, : 293 - 304
  • [6] Reliability-Aware Online Scheduling for DNN Inference Tasks in Mobile-Edge Computing
    Ma, Huirong
    Li, Rui
    Zhang, Xiaoxi
    Zhou, Zhi
    Chen, Xu
    IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (13) : 11453 - 11464
  • [7] Accelerating DNN Inference With Reliability Guarantee in Vehicular Edge Computing
    Liu, Kai
    Liu, Chunhui
    Yan, Guozhi
    Lee, Victor C. S.
    Cao, Jiannong
    IEEE-ACM TRANSACTIONS ON NETWORKING, 2023, 31 (06) : 3238 - 3253
  • [8] DNN Inference Acceleration with Partitioning and Early Exiting in Edge Computing
    Li, Chao
    Xu, Hongli
    Xu, Yang
    Wang, Zhiyuan
    Huang, Liusheng
    WIRELESS ALGORITHMS, SYSTEMS, AND APPLICATIONS, WASA 2021, PT I, 2021, 12937 : 465 - 478
  • [9] Irina: Accelerating DNN Inference with Efficient Online Scheduling
    Wu, Xiaorui
    Xu, Hong
    Wang, Yi
    PROCEEDINGS OF 2020 4TH ASIA-PACIFIC WORKSHOP ON NETWORKING, APNET 2020, 2020, : 36 - 43
  • [10] DNNShifter: An efficient DNN pruning system for edge computing
    Eccles, Bailey J.
    Rodgers, Philip
    Kilpatrick, Peter
    Spence, Ivor
    Varghese, Blesson
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 152 : 43 - 54