GossipFL: A Decentralized Federated Learning Framework With Sparsified and Adaptive Communication

被引:66
作者
Tang, Zhenheng [1 ]
Shi, Shaohuai [2 ]
Li, Bo [3 ]
Chu, Xiaowen [4 ]
机构
[1] Hong Kong Baptist Univ, Hong Kong, Peoples R China
[2] Harbin Inst Technol, Sch Comp Sci & Technol, Shenzhen, Peoples R China
[3] Hong Kong Univ Sci & Technol, Dept Comp Sci & Engn, Hong Kong, Peoples R China
[4] Hong Kong Univ Sci & Technol Guangzhou, Data Sci & Analyt Thrust, Guangzhou, Peoples R China
关键词
Deep learning; federated learning; communication efficiency;
D O I
10.1109/TPDS.2022.3230938
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Recently, federated learning (FL) techniques have enabled multiple users to train machine learning models collaboratively without data sharing. However, existing FL algorithms suffer from the communication bottleneck due to network bandwidth pressure and/or low bandwidth utilization of the participating clients in both centralized and decentralized architectures. To deal with the communication problem while preserving the convergence performance, we introduce a communication-efficient decentralized FL framework GossipFL. In GossipFL, we 1) design a novel sparsification algorithm to enable that each client only needs to communicate with one peer with a highly sparsified model, and 2) propose a new and novel gossip matrix generation algorithm that can better utilize the bandwidth resources while preserving the convergence property. We also theoretically prove that GossipFL has convergence guarantees. We conduct experiments with three convolutional neural networks on two datasets (IID and non-IID) under two distributed environments (14 clients and 100 clients) to verify the effectiveness of GossipFL. Experimental results show that GossipFL takes less communication traffic for 38.5% and less communication time for 49.8% than state-of-the-art solutions while achieving comparative model accuracy.
引用
收藏
页码:909 / 922
页数:14
相关论文
共 53 条
[1]  
Alistarh D, 2017, ADV NEUR IN, V30
[2]  
[Anonymous], 1985, Algorithmic Graph Theory
[3]   Decentralised Learning in Federated Deployment Environments: A System-Level Survey [J].
Bellavista, Paolo ;
Foschini, Luca ;
Mora, Alessio .
ACM COMPUTING SURVEYS, 2021, 54 (01)
[4]   Distributed optimization for deep learning with gossip exchange [J].
Blot, Michael ;
Picard, David ;
Thome, Nicolas ;
Cord, Matthieu .
NEUROCOMPUTING, 2019, 330 :287-296
[5]  
Bonawitz K, 2019, Arxiv, DOI arXiv:1902.01046
[6]  
Boyd S, 2005, IEEE INFOCOM SER, P1653
[7]   Randomized gossip algorithms [J].
Boyd, Stephen ;
Ghosh, Arpita ;
Prabhakar, Balaji ;
Shah, Devavrat .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2006, 52 (06) :2508-2530
[8]  
Chen CY, 2018, AAAI CONF ARTIF INTE, P2827
[9]  
Daily J, 2018, Arxiv, DOI [arXiv:1803.05880, DOI 10.48550/ARXIV.1803.05880]
[10]   Accelerating Gossip-Based Deep Learning in Heterogeneous Edge Computing Platforms [J].
Han, Rui ;
Li, Shilin ;
Wang, Xiangwei ;
Liu, Chi Harold ;
Xin, Gaofeng ;
Chen, Lydia Y. .
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2021, 32 (07) :1591-1602