Optimizing Resource Allocation for Joint AI Model Training and Task Inference in Edge Intelligence Systems

被引:14
作者
Li, Xian [1 ]
Bi, Suzhi [1 ,2 ]
Wang, Hui [1 ]
机构
[1] Shenzhen Univ, Coll Elect & Informat Engn, Shenzhen 518060, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518066, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Artificial intelligence; Training; Data models; Computational modeling; Resource management; Energy consumption; Edge intelligence; distributed training; resource allocation; alternating direction method of multipliers;
D O I
10.1109/LWC.2020.3036852
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This letter considers an edge intelligence system where multiple end users (EUs) collaboratively train an artificial intelligence (AI) model under the coordination of an edge server (ES) and the ES in return assists the AI inference task computation of EUs. Aiming at minimizing the energy consumption and execution latency of the EUs, we jointly consider the model training and task inference processes to optimize the local CPU frequency and task splitting ratio (i.e., the portion of task executed at the ES) of each EU, and the system bandwidth allocation. In particular, each task splitting ratio is correlated to a binary decision that represents whether downloading the trained AI model for local task inference. The problem is formulated as a hard mixed integer non-linear programming (MINLP). To tackle the combinatorial binary decisions, we propose a decomposition-oriented method by leveraging the ADMM (alternating direction method of multipliers) technique, whereby the primal MINLP problem is decomposed into multiple parallel sub-problems that can be efficiently handled. The proposed method enjoys linear complexity with the network size and simulation results show that it achieves near-optimal performance (less than 3.18% optimality gap), which significantly outperforms the considered benchmark algorithms.
引用
收藏
页码:532 / 536
页数:5
相关论文
共 14 条
  • [1] Computation Rate Maximization for Wireless Powered Mobile-Edge Computing With Binary Computation Offloading
    Bi, Suzhi
    Zhang, Ying Jun
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2018, 17 (06) : 4177 - 4190
  • [2] Distributed optimization and statistical learning via the alternating direction method of multipliers
    Boyd S.
    Parikh N.
    Chu E.
    Peleato B.
    Eckstein J.
    [J]. Foundations and Trends in Machine Learning, 2010, 3 (01): : 1 - 122
  • [3] Chen M., 2019, ARXIV190907972
  • [4] Chenoweth JM, 2016, FLA MUS NAT HIST-RIP, P1
  • [5] An Overview of Dynamic-Linearization-Based Data-Driven Control and Applications
    Hou, Zhongsheng
    Chi, Ronghu
    Gao, Huijun
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2017, 64 (05) : 4076 - 4090
  • [6] Joint Offloading and Computing Optimization in Wireless Powered Mobile-Edge Computing Systems
    Wang, Feng
    Xu, Jie
    Wang, Xin
    Cui, Shuguang
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2018, 17 (03) : 1784 - 1797
  • [7] Wang SQ, 2018, IEEE INFOCOM SER, P63, DOI 10.1109/INFOCOM.2018.8486403
  • [8] Convergence of Edge Computing and Deep Learning: A Comprehensive Survey
    Wang, Xiaofei
    Han, Yiwen
    Leung, Victor C. M.
    Niyato, Dusit
    Yan, Xueqiang
    Chen, Xu
    [J]. IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2020, 22 (02): : 869 - 904
  • [9] Mobile-Edge Computing: Partial Computation Offloading Using Dynamic Voltage Scaling
    Wang, Yanting
    Sheng, Min
    Wang, Xijun
    Wang, Liang
    Li, Jiandong
    [J]. IEEE TRANSACTIONS ON COMMUNICATIONS, 2016, 64 (10) : 4268 - 4282
  • [10] Optimal Task Offloading and Resource Allocation in Mobile-Edge Computing With Inter-User Task Dependency
    Yan, Jia
    Bi, Suzhi
    Zhang, Ying Jun
    Tao, Meixia
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2020, 19 (01) : 235 - 250