Communication-Efficient Federated Learning via Regularized Sparse Random Networks

被引:0
作者
Mestoukirdi, Mohamad [1 ]
Esrafilian, Omid [2 ]
Gesbert, David [2 ]
Li, Qianrui [1 ,3 ,4 ]
Gresset, Nicolas [1 ]
机构
[1] Mitsubishi Elect R& Ctr Europe, F-35700 Rennes, France
[2] Eurecom, Commun Syst Dept, F-06904 Sophia Antipolis, France
[3] CICT Mobile Commun Technol Co Ltd, Beijing 100832, Peoples R China
[4] China Acad Telecommun Technol CATT, State Key Lab Wireless Mobile Commun, Beijing 100191, Peoples R China
关键词
Vectors; Training; Entropy; Servers; Federated learning; Stochastic processes; Memory management; sparse random networks; unstructured sparsity;
D O I
10.1109/LCOMM.2024.3397250
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
This work presents a new method for enhancing communication efficiency in stochastic Federated Learning that trains over-parameterized random networks. In this setting, a binary mask is optimized instead of the model weights, which are kept fixed. The mask characterizes a sparse sub-network that is able to generalize as good as a smaller target network. Importantly, sparse binary masks are exchanged rather than the floating point weights in traditional federated learning, reducing communication cost to at most 1 bit per parameter (Bpp). We show that previous state of the art stochastic methods fail to find sparse networks that can reduce the communication and storage overhead using consistent loss objectives. To address this, we propose adding a regularization term to local objectives that acts as a proxy of the transmitted masks entropy, therefore encouraging sparser solutions by eliminating redundant features across sub-networks. Extensive empirical experiments demonstrate significant improvements in communication and memory efficiency of up to five magnitudes compared to the literature, with minimal performance degradation in validation accuracy in some instances.
引用
收藏
页码:1574 / 1578
页数:5
相关论文
共 12 条
[1]  
Alistarh D, 2017, Arxiv, DOI arXiv:1610.02132
[2]  
Bernstein J, 2018, Arxiv, DOI arXiv:1802.04434
[3]  
Brendan McMahan H., CORR
[4]  
Deng Li., 2012, Signal Processing Magazine, IEEE, V29, P141, DOI [DOI 10.1109/MSP.2012.2211477, 10.1109/msp.2012.2211477]
[5]  
Isik B, 2022, Arxiv, DOI arXiv:2209.15328
[6]   FedMask: Joint Computation and Communication-Efficient Personalized Federated Learning via Heterogeneous Masking [J].
Li, Ang ;
Sun, Jingwei ;
Zeng, Xiao ;
Zhang, Mi ;
Li, Hai ;
Chen, Yiran .
PROCEEDINGS OF THE 2021 THE 19TH ACM CONFERENCE ON EMBEDDED NETWORKED SENSOR SYSTEMS, SENSYS 2021, 2021, :42-55
[7]  
Lueker GS, 1998, RANDOM STRUCT ALGOR, V12, P51, DOI 10.1002/(SICI)1098-2418(199801)12:1<51::AID-RSA3>3.0.CO
[8]  
2-S
[9]  
Pensia A.Pensia, 2020, arXiv, V12
[10]  
Ramanujan V, 2020, Arxiv, DOI [arXiv:1911.13299, DOI 10.48550/ARXIV.1911.13299, 10.48550/ARXIV.1911.13299]