Multi-fine-grained DNNs Partition and Offloading over Fog Computing Networks

被引:2
作者
Fan, Xuwei [1 ]
Cheng, Zhipeng [1 ]
Liwang, Minghui [1 ]
Chen, Ning [1 ]
Huang, Lianfen [1 ]
Wang, Xianbin [2 ]
机构
[1] Xiamen Univ, Sch Informat, Xiamen, Peoples R China
[2] Univ Western Ontario, Dept Elect & Comp Engn, London, ON, Canada
来源
2023 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS, ICC WORKSHOPS | 2023年
基金
中国国家自然科学基金;
关键词
DNNs inference; DNNs partition; DNNs offloading; Multiagent Hybrid Actions Deep Deterministic Policy Gradient (MHADDPG); Fog computing; INFERENCE;
D O I
10.1109/ICCWORKSHOPS57953.2023.10283497
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks (DNNs) have facilitated commendable performance in signal processing, thanks to their superior functions in feature extraction and abstraction representation. However, limited computing capacity of Internet of Things (IoT) devices imposes challenges to support resource-intensive DNNs inference with low latency and quality of service requirements. Benefiting from the pervasive wireless connectivity, offloading partial DNNs to fog nodes (FNs) becomes a viable solution for alleviating resource shortage and improving time and energy efficiency. This paper investigates a novel multi-fine-grained DNNs partition and offloading strategy over fog computing networks. A Multiagent Hybrid Actions Deep Deterministic Policy Gradient (MHADDPG)-based algorithm is proposed to maximize the long-term system utility, upon considering the DNNs execution delay and energy consumption of participating devices and FNs. Comprehensive simulations demonstrate that the proposed solution significantly reduces the average inference latency by 55.15%-67.08%, while saving energy by 44.39%-57.56%, for three widely adopted DNNs.
引用
收藏
页码:599 / 604
页数:6
相关论文
共 11 条
  • [1] Joint Task Offloading and Resource Allocation for Mobile Edge Computing in Ultra-Dense Network
    Cheng, Zhipeng
    Min, Minghui
    Gao, Zhibin
    Huang, Lianfen
    [J]. 2020 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2020,
  • [2] Mohammed T, 2020, IEEE INFOCOM SER, P854, DOI [10.1109/INFOCOM41043.2020.9155237, 10.1109/infocom41043.2020.9155237]
  • [3] Efficient Processing of Deep Neural Networks: A Tutorial and Survey
    Sze, Vivienne
    Chen, Yu-Hsin
    Yang, Tien-Ju
    Emer, Joel S.
    [J]. PROCEEDINGS OF THE IEEE, 2017, 105 (12) : 2295 - 2329
  • [4] Tuli S., 2021, arXiv
  • [5] Joint Optimization of Caching, Computing, and Radio Resources for Fog-Enabled IoT Using Natural Actor-Critic Deep Reinforcement Learning
    Wei, Yifei
    Yu, F. Richard
    Song, Mei
    Han, Zhu
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2019, 6 (02) : 2061 - 2073
  • [6] FedAdapt: Adaptive Offloading for IoT Devices in Federated Learning
    Wu, Di
    Ullah, Rehmat
    Harvey, Paul
    Kilpatrick, Peter
    Spence, Ivor
    Varghese, Blesson
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (21) : 20889 - 20901
  • [7] Energy-Aware Inference Offloading for DNN-Driven Applications in Mobile Edge Clouds
    Xu, Zichuan
    Zhao, Liqian
    Liang, Weifa
    Rana, Omer F.
    Zhou, Pan
    Xia, Qiufen
    Xu, Wenzheng
    Wu, Guowei
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2021, 32 (04) : 799 - 814
  • [8] DDPQN: An Efficient DNN Offloading Strategy in Local-Edge-Cloud Collaborative Environments
    Xue, Min
    Wu, Huaming
    Peng, Guang
    Wolter, Katinka
    [J]. IEEE TRANSACTIONS ON SERVICES COMPUTING, 2022, 15 (02) : 640 - 655
  • [9] CoEdge: Cooperative DNN Inference With Adaptive Workload Partitioning Over Heterogeneous Edge Devices
    Zeng, Liekang
    Chen, Xu
    Zhou, Zhi
    Yang, Lei
    Zhang, Junshan
    [J]. IEEE-ACM TRANSACTIONS ON NETWORKING, 2021, 29 (02) : 595 - 608
  • [10] Energy-Optimal Mobile Cloud Computing under Stochastic Wireless Channel
    Zhang, Weiwen
    Wen, Yonggang
    Guan, Kyle
    Kilper, Dan
    Luo, Haiyun
    Wu, Dapeng Oliver
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2013, 12 (09) : 4569 - 4581