Distributed Deep Neural Network Training on Edge Devices

被引:9
作者
Benditkis, Daniel [1 ]
Keren, Aviv [1 ]
Mor-Yosef, Liron [1 ]
Avidor, Tomer [1 ]
Shoham, Neta [1 ]
Tal-Israel, Nadav [1 ]
机构
[1] Edgify, Ramat Gan, Israel
来源
SEC'19: PROCEEDINGS OF THE 4TH ACM/IEEE SYMPOSIUM ON EDGE COMPUTING | 2019年
关键词
edge device; neural network; deep learning; federated learning; large batch; communication compression;
D O I
10.1145/3318216.3363324
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Neural Network (Deep Learning) models have been traditionally trained on dedicated servers, after collecting data from various edge devices and sending them to the server. In recent years new methodologies have emerged for training models in a distributed manner over edge devices, keeping the data on the devices themselves. This allows for better data privacy and reduces the training costs. One of the main challenges for such methodologies is reducing the communication costs to and mainly from the edge devices. In this work we compare the two main methodologies used for distributed edge training: Federated Learning and Large Batch Training. For each of the methodologies we examine their convergence rates, communication costs, and final model performance. In addition, we present two techniques for compressing the communication between the edge devices, and examine their suitability for each one of the training methodologies.
引用
收藏
页码:304 / 306
页数:3
相关论文
共 6 条
[1]  
[Anonymous], 2017, ICLR
[2]  
Goyal P., ACCURATE LARGE MINIB
[3]  
He K., 2016, 2016 IEEE C COMPUTER
[4]  
Konecny J., Federated learning: strategies for improving communication efficiency
[5]  
Lin Yujun, 2017, ARXIV PREPRINT ARXIV
[6]  
McMahan H. B., COMMUNICATION EFFICI