FedQNN: A Computation-Communication-Efficient Federated Learning Framework for IoT With Low-Bitwidth Neural Network Quantization

被引:23
作者
Ji, Yu [1 ,2 ]
Chen, Lan [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Microelect, Beijing 100029, Peoples R China
[2] Beijing Municipal Commiss Sci & Technol, Beijing Key Lab Three Dimens & Nanometer Integrat, Beijing 100029, Peoples R China
关键词
Deep learning; edge computing; federated learning (FL); neural network quantization;
D O I
10.1109/JIOT.2022.3213650
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL) allows participants to train deep learning models collaboratively without disclosing their data to the server or any other participants, providing excellent value in the field of privacy-sensitive IoT. However, this distributed training paradigm requires clients to perform intensive computation for many iterations, which may exceed the capability of a typical IoT terminal with limited processing power, storage capacity, and energy budget. Heavy communication between the server and clients may also result in intolerant bandwidth requirements and energy consumption for many IoT systems. In this article, we introduce the FedQNN, a computation-communication-efficient FL framework for IoT scenarios. It is the first work that integrates ultralow-bitwidth quantization into the FL environment, allowing clients to perform lightweight fix-point computation efficiently with less power. Furthermore, both upstream and downstream data are significantly compressed for more efficient communication using a combination of sparsification and quantization strategies. We performed extensive experiments on a variety of data sets and models while comparing with other frameworks, and the results demonstrate that the proposed method can save up to 90% of our clients' computational energy, reduce model sizes by 30+ times, and significantly compress both communication bandwidth and transmitted data size while maintaining reasonable accuracy. The robustness against the non-independent and identically distributed (I.I.D.) condition is also validated.
引用
收藏
页码:2494 / 2507
页数:14
相关论文
共 65 条
[1]  
Al Bahou A, 2018, PROC IEEE COOL CHIPS
[2]  
Alistarh D, 2017, ADV NEUR IN, V30
[3]  
Allen-Zhu Z, 2019, PR MACH LEARN RES, V97
[4]  
Ando K, 2017, SYMP VLSI CIRCUITS, pC24, DOI 10.23919/VLSIC.2017.8008533
[5]   YodaNN1 : An Ultra-Low Power Convolutional Neural Network Accelerator Based on Binary Weights [J].
Andri, Renzo ;
Cavigelli, Lukas ;
Rossi, Davide ;
Benini, Luca .
2016 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI), 2016, :236-241
[6]  
Banner Ron, 2018, Advances in neural information processing systems, V31
[7]  
Bernstein J, 2018, PR MACH LEARN RES, V80
[8]   Large-Scale Machine Learning with Stochastic Gradient Descent [J].
Bottou, Leon .
COMPSTAT'2010: 19TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL STATISTICS, 2010, :177-186
[9]  
Chen ZL, 2017, IEEE INT VEH SYM, P1856, DOI 10.1109/IVS.2017.7995975
[10]   Distributed Differential Privacy via Shuffling [J].
Cheu, Albert ;
Smith, Adam ;
Ullman, Jonathan ;
Zeber, David ;
Zhilyaev, Maxim .
ADVANCES IN CRYPTOLOGY - EUROCRYPT 2019, PT I, 2019, 11476 :375-403