Communication-Efficient Federated Learning for Wireless Edge Intelligence in IoT

被引:311
作者
Mills, Jed [1 ]
Hu, Jia [1 ]
Min, Geyong [1 ]
机构
[1] Univ Exeter, Coll Engn Maths & Phys Sci, Exeter EX4 4QJ, Devon, England
基金
英国工程与自然科学研究理事会;
关键词
Compression; distributed computing; edge computing; federated learning (FL); Internet of Things (IoT); INTERNET;
D O I
10.1109/JIOT.2019.2956615
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The rapidly expanding number of Internet of Things (IoT) devices is generating huge quantities of data, but public concern over data privacy means users are apprehensive to send data to a central server for machine learning (ML) purposes. The easily changed behaviors of edge infrastructure that software-defined networking (SDN) provides makes it possible to collate IoT data at edge servers and gateways, where federated learning (FL) can be performed: building a central model without uploading data to the server. FedAvg is an FL algorithm which has been the subject of much study, however, it suffers from a large number of rounds to convergence with non-independent identically distributed (non-IID) client data sets and high communication costs per round. We propose adapting FedAvg to use a distributed form of Adam optimization, greatly reducing the number of rounds to convergence, along with the novel compression techniques, to produce communication-efficient FedAvg (CE-FedAvg). We perform extensive experiments with the MNIST/CIFAR-10 data sets, IID/non-IID client data, varying numbers of clients, client participation rates, and compression rates. These show that CE-FedAvg can converge to a target accuracy in up to 6x less rounds than similarly compressed FedAvg, while uploading up to 3x less data, and is more robust to aggressive compression. Experiments on an edge-computing-like testbed using Raspberry Pi clients also show that CE-FedAvg is able to reach a target accuracy in up to 1.7x less real time than FedAvg.
引用
收藏
页码:5986 / 5994
页数:9
相关论文
共 30 条
[1]  
Abadi Martin, 2016, Proceedings of OSDI '16: 12th USENIX Symposium on Operating Systems Design and Implementation. OSDI '16, P265
[2]  
[Anonymous], 2016, CORR
[3]  
[Anonymous], 2017, Collaborative machine learning without centralized training data
[4]   Strategies for Re-training a Pruned Neural Network in an Edge Computing Paradigm [J].
Chandakkar, Parag S. ;
Li, Yikang ;
Ding, Pak Lun Kevin ;
Li, Baoxin .
2017 IEEE 1ST INTERNATIONAL CONFERENCE ON EDGE COMPUTING (IEEE EDGE), 2017, :244-247
[5]  
Chen Y., 2019, ARXIV
[6]  
Duchi J, 2011, J MACH LEARN RES, V12, P2121
[7]   RUN-LENGTH ENCODINGS [J].
GOLOMB, SW .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1966, 12 (03) :399-+
[8]   A New Learning Automata-Based Pruning Method to Train Deep Neural Networks [J].
Guo, Haonan ;
Li, Shenghong ;
Li, Bin ;
Ma, Yinghua ;
Ren, Xudie .
IEEE INTERNET OF THINGS JOURNAL, 2018, 5 (05) :3263-3269
[9]  
Hard A., 2018, CoRR
[10]  
King DB, 2015, ACS SYM SER, V1214, P1, DOI 10.1021/bk-2015-1214.ch001