ARFL: Adaptive and Robust Federated Learning

被引:4
作者
Uddin, Md Palash [1 ]
Xiang, Yong [1 ]
Cai, Borui [1 ]
Lu, Xuequan [2 ]
Yearwood, John [1 ]
Gao, Longxiang [3 ,4 ]
机构
[1] Deakin Univ, Sch Informat Technol, Geelong, Vic 3220, Australia
[2] La Trobe Univ, Bundoora, Vic 3086, Australia
[3] Qilu Univ Technol, Shandong Acad Sci, Jinan 250316, Shandong, Peoples R China
[4] Nat Supercomp Ctr Jinan, Shandong Comp Sci Ctr, Jinan 250101, Shandong, Peoples R China
基金
澳大利亚研究理事会;
关键词
Distributed learning; federated learning; parallel optimization; communication overhead; adaptive workload; adaptive step size; proximal term; robust aggregation;
D O I
10.1109/TMC.2023.3310248
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated Learning (FL) is a machine learning technique that enables multiple local clients holding individual datasets to collaboratively train a model, without exchanging the clients' datasets. Conventional FL approaches often assign a fixed workload (local epoch) and step size (learning rate) to the clients during the client-side local model training and utilize all collaborating trained models' parameters evenly during the server-side global model aggregation. Consequently, they frequently experience problems with data heterogeneity and high communication costs. In this paper, we propose a novel FL approach to mitigate the above problems. On the client side, we propose an adaptive model update approach that optimally allocates a needful number of local epochs and dynamically adjusts the learning rate to train the local model and regularizes the conventional objective function by adding a proximal term to it. On the server side, we propose a robust model aggregation strategy that potentially supplants the local outlier updates (models' weights) prior to the aggregation. We provide the theoretical convergence results and perform extensive experiments on different data setups over the MNIST, CIFAR-10, and Shakespeare datasets, which manifest that our FL scheme surpasses the baselines in terms of communication speedup, test-set performance, and global convergence.
引用
收藏
页码:5401 / 5417
页数:17
相关论文
共 42 条
  • [1] FedPacket: A Federated Learning Approach to Mobile Packet Classification
    Bakopoulou, Evita
    Tillman, Balint
    Markopoulou, Athina
    [J]. IEEE TRANSACTIONS ON MOBILE COMPUTING, 2022, 21 (10) : 3609 - 3628
  • [2] Bonawitz K., 2016, ARXIV
  • [3] Bonawitz K., 2019, PROC MACH LEARN SYST, V1, P374
  • [4] Practical Secure Aggregation for Privacy-Preserving Machine Learning
    Bonawitz, Keith
    Ivanov, Vladimir
    Kreuter, Ben
    Marcedone, Antonio
    McMahan, H. Brendan
    Patel, Sarvar
    Ramage, Daniel
    Segal, Aaron
    Seth, Karn
    [J]. CCS'17: PROCEEDINGS OF THE 2017 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2017, : 1175 - 1191
  • [5] Dean J., 2012, NIPS
  • [6] ORDER AND STEPSIZE CONTROL IN EXTRAPOLATION METHODS
    DEUFLHARD, P
    [J]. NUMERISCHE MATHEMATIK, 1983, 41 (03) : 399 - 422
  • [7] Self-Balancing Federated Learning With Global Imbalanced Data in Mobile Systems
    Duan, Moming
    Liu, Duo
    Chen, Xianzhang
    Liu, Renping
    Tan, Yujuan
    Liang, Liang
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2021, 32 (01) : 59 - 71
  • [8] Fausett L., 1994, FUNDAMENTALS NEURAL
  • [9] Golmant N, 2018, Arxiv, DOI [arXiv:1811.12941, 10.48550/arXiv.1811.12941]
  • [10] Patient clustering improves efficiency of federated machine learning to predict mortality and hospital stay time using distributed electronic medical records
    Huang, Li
    Shea, Andrew L.
    Qian, Huining
    Masurkar, Aditya
    Deng, Hao
    Liu, Dianbo
    [J]. JOURNAL OF BIOMEDICAL INFORMATICS, 2019, 99