ARFL: Adaptive and Robust Federated Learning

被引:4
作者
Uddin, Md Palash [1 ]
Xiang, Yong [1 ]
Cai, Borui [1 ]
Lu, Xuequan [2 ]
Yearwood, John [1 ]
Gao, Longxiang [3 ,4 ]
机构
[1] Deakin Univ, Sch Informat Technol, Geelong, Vic 3220, Australia
[2] La Trobe Univ, Bundoora, Vic 3086, Australia
[3] Qilu Univ Technol, Shandong Acad Sci, Jinan 250316, Shandong, Peoples R China
[4] Nat Supercomp Ctr Jinan, Shandong Comp Sci Ctr, Jinan 250101, Shandong, Peoples R China
基金
澳大利亚研究理事会;
关键词
Distributed learning; federated learning; parallel optimization; communication overhead; adaptive workload; adaptive step size; proximal term; robust aggregation;
D O I
10.1109/TMC.2023.3310248
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated Learning (FL) is a machine learning technique that enables multiple local clients holding individual datasets to collaboratively train a model, without exchanging the clients' datasets. Conventional FL approaches often assign a fixed workload (local epoch) and step size (learning rate) to the clients during the client-side local model training and utilize all collaborating trained models' parameters evenly during the server-side global model aggregation. Consequently, they frequently experience problems with data heterogeneity and high communication costs. In this paper, we propose a novel FL approach to mitigate the above problems. On the client side, we propose an adaptive model update approach that optimally allocates a needful number of local epochs and dynamically adjusts the learning rate to train the local model and regularizes the conventional objective function by adding a proximal term to it. On the server side, we propose a robust model aggregation strategy that potentially supplants the local outlier updates (models' weights) prior to the aggregation. We provide the theoretical convergence results and perform extensive experiments on different data setups over the MNIST, CIFAR-10, and Shakespeare datasets, which manifest that our FL scheme surpasses the baselines in terms of communication speedup, test-set performance, and global convergence.
引用
收藏
页码:5401 / 5417
页数:17
相关论文
共 42 条
  • [11] Distillation-Based Semi-Supervised Federated Learning for Communication-Efficient Collaborative Training With Non-IID Private Data
    Itahara, Sohei
    Nishio, Takayuki
    Koda, Yusuke
    Morikura, Masahiro
    Yamamoto, Koji
    [J]. IEEE TRANSACTIONS ON MOBILE COMPUTING, 2023, 22 (01) : 191 - 205
  • [12] Johnson R., 2013, Advances in Neural Information Processing Systems, P315
  • [13] Konecny J., 2016, Comput. Res. Repository, P1
  • [14] Konecny J., 2016, PROC INT C NEURAL IN, P1
  • [15] Koskela A, 2020, PR MACH LEARN RES, V108, P2465
  • [16] Li T., 2020, PROC MACH LEARN SYST, P1
  • [17] Li X., 2020, PROC 8 INT C LEARN R, P1
  • [18] Accelerating Federated Learning via Momentum Gradient Descent
    Liu, Wei
    Chen, Li
    Chen, Yunfei
    Zhang, Wenyi
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2020, 31 (08) : 1754 - 1766
  • [19] Differentially Private Asynchronous Federated Learning for Mobile Edge Computing in Urban Informatics
    Lu, Yunlong
    Huang, Xiaohong
    Dai, Yueyue
    Maharjan, Sabita
    Zhang, Yan
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2020, 16 (03) : 2134 - 2143
  • [20] Adaptive Batch Size for Federated Learning in Resource-Constrained Edge Computing
    Ma, Zhenguo
    Xu, Yang
    Xu, Hongli
    Meng, Zeyu
    Huang, Liusheng
    Xue, Yinxing
    [J]. IEEE TRANSACTIONS ON MOBILE COMPUTING, 2023, 22 (01) : 37 - 53