AFedAvg: communication-efficient federated learning aggregation with adaptive communication frequency and gradient sparse

被引:5
作者
Li, Yanbin [1 ]
He, Ziming [1 ]
Gu, Xingjian [1 ]
Xu, Huanliang [1 ]
Ren, Shougang [1 ]
机构
[1] Nanjing Agr Univ, Coll Artificial Intelligence, Nanjing, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
Federated learning; communication cost; gradient sparse; communication frequency; OPTIMIZATION;
D O I
10.1080/0952813X.2022.2079730
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Federated learning enables a large number of clients (such as edge computing devices) to learn a model jointly without data sharing. However, the high amount of communication of the federated learning aggregation algorithm hinders the realisation of artificial intelligence in the last mile. Although FederatedAveraging (FedAvg) is the leading algorithm, its communication cost is still high. The method of communication delay and gradient sparse can reduce the communication cost, but there is no previous work to analyse the relationship and common effects of these two dimensions. Aiming at the problems that federated learning communication is expensive and it has become a training bottleneck, we improve the FedAvg algorithm and propose an adaptive communication frequency FederatedAveraging algorithm (AFedAvg). The gradient sparse operation in the algorithm reduces the quantity of parameters for a single communication, while the communication delay operation allows training to converge faster and obtain smaller losses. The number of sparse parameters is used to select the communication frequency of next round dynamically. Experimental results prove that, the AFedAvg algorithm is superior to the FedAvg and its variants in terms of communication cost. It achieves 2.4X-23.1X communication compression in different data distributions with minimal communication rounds required by the algorithm to converge.
引用
收藏
页码:47 / 69
页数:23
相关论文
共 40 条
  • [1] Ahmed A., 2012, P 5 ACM INT C WEB SE, P123, DOI 10.1145/2124295.2124312
  • [2] Aji A. F., 2017, P 2017 C EMPIRICAL M, P440, DOI DOI 10.18653/V1/D17-1045
  • [3] [Anonymous], 2015, ARXIV151106051
  • [4] [Anonymous], ACM Trans. Intell. Syst. Technol. TIST, DOI [DOI 10.1145/3298981, 10.1145/3298981]
  • [5] A Joint Learning and Communications Framework for Federated Learning Over Wireless Networks
    Chen, Mingzhe
    Yang, Zhaohui
    Saad, Walid
    Yin, Changchuan
    Poor, H. Vincent
    Cui, Shuguang
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (01) : 269 - 283
  • [6] Dean J., 2012, PROC NEURIPS 12, P1232
  • [7] Dekel O, 2012, J MACH LEARN RES, V13, P165
  • [8] Elephant detection using boundary sense deep learning (BSDL) architecture
    Dhanaraj, Jerline Sheebha Anni
    Sangaiah, Arun Kumar
    [J]. JOURNAL OF EXPERIMENTAL & THEORETICAL ARTIFICIAL INTELLIGENCE, 2021, 33 (04) : 561 - 576
  • [9] Dryden N, 2016, PROCEEDINGS OF 2016 2ND WORKSHOP ON MACHINE LEARNING IN HPC ENVIRONMENTS (MLHPC), P1, DOI [10.1109/MLHPC.2016.4, 10.1109/MLHPC.2016.004]
  • [10] Cluster Pruning: An Efficient Filter Pruning Method for Edge AI Vision Applications
    Gamanayake, Chinthaka
    Jayasinghe, Lahiru
    Ng, Benny Kai Kiat
    Yuen, Chau
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (04) : 802 - 816