Joint Task Offloading and Resource Allocation for Quality-Aware Edge-Assisted Machine Learning Task Inference

被引:5
作者
Fan, Wenhao [1 ,2 ]
Chen, Zeyu [1 ,2 ]
Hao, Zhibo [1 ,2 ]
Wu, Fan [1 ,2 ]
Liu, Yuan'an [1 ,2 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Elect Engn, Beijing 100876, Peoples R China
[2] Beijing Univ Posts & Telecommun, Beijing Key Lab Work Safety Intelligent Monitoring, Beijing 100876, Peoples R China
基金
北京市自然科学基金;
关键词
Task analysis; Servers; Computational modeling; Optimization; Delays; Resource management; Data models; Task offloading; data quality; edge computing; edge intelligence; inference;
D O I
10.1109/TVT.2023.3235520
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Edge computing is essential to enhance delay-sensitive and computation-intensive machine learning (ML) task inference services. Quality of inference results, which is mainly impacted by the task data and ML models, is an important indicator impacting the system performance. In this paper, we consider a quality-aware edge-assisted ML task inference scenario and propose a resource management scheme to minimize the total task processing delay while guaranteeing the stability of all the task queues and the inference accuracy requirements of all the tasks. In our scheme, the task offloading, task data adjustment, computing resource allocation, and wireless channel allocation are jointly optimized. The Lyapunov optimization technique is adopted to transform the original optimization problem into a deterministic problem for each time slot. Considering the high complexity of the optimization problem, we design an algorithm that decomposes the problem into a task offloading and channel allocation (TOCA) sub-problem, a task data adjustment sub-problem, and a computing resource allocation sub-problem, and then solves them iteratively. A low-complexity heuristic algorithm is also designed to solve the TOCA sub-problem efficiently. Extensive simulations are conducted by varying different crucial parameters. The results demonstrate the superiority of our scheme in comparison with 4 other schemes.
引用
收藏
页码:6739 / 6752
页数:14
相关论文
共 32 条
[1]  
[Anonymous], DAGM 2007
[2]   402 Gb/s PAM-8 IM/DD O-Band EML Transmission [J].
Bin Hossain, Md Sabbir ;
Wei, Jinlong ;
Pittala, Fabio ;
Stojanovic, Nebojsa ;
Calabro, Stefano ;
Rahman, Talha ;
Bocherer, Georg ;
Wettlin, Tom ;
Xie, Changsong ;
Kuschnerov, Maxim ;
Pachnicke, Stephan .
2021 EUROPEAN CONFERENCE ON OPTICAL COMMUNICATION (ECOC), 2021,
[3]  
Bochkovskiy A, 2020, Arxiv, DOI arXiv:2004.10934
[4]  
Boyd SP., 2004, Convex optimization, DOI 10.1017/CBO9780511804441
[5]   Adaptive Fog Configuration for the Industrial Internet of Things [J].
Chen, Lixing ;
Zhou, Pan ;
Gao, Liang ;
Xu, Jie .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2018, 14 (10) :4656-4664
[6]   Compressing Deep Model With Pruning and Tucker Decomposition for Smart Embedded Systems [J].
Dai, Cheng ;
Liu, Xingang ;
Cheng, Hongqiang ;
Yang, Laurence T. ;
Deen, M. Jamal .
IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (16) :14490-14500
[7]   Edge Intelligence: The Confluence of Edge Computing and Artificial Intelligence [J].
Deng, Shuiguang ;
Zhao, Hailiang ;
Fang, Weijia ;
Yin, Jianwei ;
Dustdar, Schahram ;
Zomaya, Albert Y. .
IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (08) :7457-7469
[8]   Q-GADMM: Quantized Group ADMM for Communication Efficient Decentralized Machine Learning [J].
Elgabli, Anis ;
Park, Jihong ;
Bedi, Amrit Singh ;
Ben Issaid, Chaouki ;
Bennis, Mehdi ;
Aggarwal, Vaneet .
IEEE TRANSACTIONS ON COMMUNICATIONS, 2021, 69 (01) :164-181
[9]   Collaborative Service Placement, Task Scheduling, and Resource Allocation for Task Offloading With Edge-Cloud Cooperation [J].
Fan, Wenhao ;
Zhao, Liang ;
Liu, Xun ;
Su, Yi ;
Li, Shenmeng ;
Wu, Fan ;
Liu, Yuan'an .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (01) :238-256
[10]   Joint Task Offloading and Resource Allocation for Accuracy-Aware Machine-Learning-Based IIoT Applications [J].
Fan, Wenhao ;
Li, Shenmeng ;
Liu, Jie ;
Su, Yi ;
Wu, Fan ;
Liu, Yuan'An .
IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (04) :3305-3321