HierTrain: Fast Hierarchical Edge AI Learning With Hybrid Parallelism in Mobile-Edge-Cloud Computing

被引:43
作者
Liu, Deyin [1 ]
Chen, Xu [1 ]
Zhou, Zhi [1 ]
Ling, Qing [1 ]
机构
[1] Sun Yat Sen Univ, Sch Data & Comp Sci, Guangzhou 510006, Peoples R China
来源
IEEE OPEN JOURNAL OF THE COMMUNICATIONS SOCIETY | 2020年 / 1卷
基金
美国国家科学基金会;
关键词
Edge AI; deep learning; fast model training; mobile-edge-cloud computing; ENERGY;
D O I
10.1109/OJCOMS.2020.2994737
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Nowadays, deep neural networks (DNNs) are the core enablers for many emerging edge AI applications. Conventional approaches for training DNNs are generally implemented at central servers or cloud centers for centralized learning, which is typically time-consuming and resource-demanding due to the transmission of a large number of data samples from the edge device to the remote cloud. To overcome these disadvantages, we consider accelerating the learning process of DNNs on the Mobile-Edge-Cloud Computing (MECC) paradigm. In this paper, we propose HierTrain, a hierarchical edge AI learning framework, which efficiently deploys the DNN training task over the hierarchical MECC architecture. We develop a novel hybrid parallelism method, which is the key to HierTrain, to adaptively assign the DNN model layers and the data samples across the three levels of the edge device, edge server and cloud center. We then formulate the problem of scheduling the DNN training tasks at both layer-granularity and sample-granularity. Solving this optimization problem enables us to achieve the minimum training time. We further implement a hardware prototype consisting of an edge device, an edge server and a cloud server, and conduct extensive experiments on it. Experimental results demonstrate that HierTrain can achieve up to 6.9x speedups compared to the cloud-based hierarchical training approach.
引用
收藏
页码:634 / 645
页数:12
相关论文
共 42 条
[1]  
Abadi M, 2016, ACM SIGPLAN NOTICES, V51, P1, DOI [10.1145/3022670.2976746, 10.1145/2951913.2976746]
[2]  
[Anonymous], 2010, PROC 9 PYTHON SCI
[3]  
[Anonymous], 2015, OPEN J CLOUD COMPUTI
[4]   Large-Scale Machine Learning with Stochastic Gradient Descent [J].
Bottou, Leon .
COMPSTAT'2010: 19TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL STATISTICS, 2010, :177-186
[5]  
Chahal K., 2018, ARXIV181109878
[6]   EXPLOITING MASSIVE D2D COLLABORATION FOR ENERGY-EFFICIENT MOBILE EDGE COMPUTING [J].
Chen, Xu ;
Pu, Lingjun ;
Gao, Lin ;
Wu, Weigang ;
Wu, Di .
IEEE WIRELESS COMMUNICATIONS, 2017, 24 (04) :64-71
[7]   High Prevalence of Assisted Injection Among Street-Involved Youth in a Canadian Setting [J].
Cheng, Tessa ;
Kerr, Thomas ;
Small, Will ;
Dong, Huiru ;
Montaner, Julio ;
Wood, Evan ;
DeBeck, Kora .
AIDS AND BEHAVIOR, 2016, 20 (02) :377-384
[8]   Deep Neural Networks for YouTube Recommendations [J].
Covington, Paul ;
Adams, Jay ;
Sargin, Emre .
PROCEEDINGS OF THE 10TH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS'16), 2016, :191-198
[9]   Bringing Transparency Design into Practice [J].
Eiband, Malin ;
Schneider, Hanna ;
Bilandzic, Mark ;
Fazekas-Con, Julian ;
Haug, Mareike ;
Hussmann, Heinrich .
IUI 2018: PROCEEDINGS OF THE 23RD INTERNATIONAL CONFERENCE ON INTELLIGENT USER INTERFACES, 2018, :211-223
[10]   JointDNN: An Efficient Training and Inference Engine for Intelligent Mobile Cloud Computing Services [J].
Eshratifar, Amir Erfan ;
Abrishami, Mohammad Saeed ;
Pedram, Massoud .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2021, 20 (02) :565-576