Generalized Federated Learning via Gradient Norm-Aware Minimization and Control Variables

被引:1
作者
Xu, Yicheng [1 ]
Ma, Wubin [1 ]
Dai, Chaofan [1 ]
Wu, Yahui [1 ]
Zhou, Haohao [1 ]
机构
[1] Natl Univ Def Technol, Coll Syst Engn, Changsha 410073, Peoples R China
基金
中国国家自然科学基金;
关键词
federated learning; client drift; distributed learning;
D O I
10.3390/math12172644
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Federated Learning (FL) is a promising distributed machine learning framework that emphasizes privacy protection. However, inconsistencies between local optimization objectives and the global objective, commonly referred to as client drift, primarily arise due to non-independently and identically distributed (Non-IID) data, multiple local training steps, and partial client participation in training. The majority of current research tackling this challenge is mainly based on the empirical risk minimization (ERM) principle, while giving little consideration to the connection between the global loss landscape and generalization capability. This study proposes FedGAM, an innovative FL algorithm that incorporates Gradient Norm-Aware Minimization (GAM) to efficiently search for a local flat landscape. FedGAM specifically modifies the client model training objective to simultaneously minimize the loss value and first-order flatness, thereby seeking flat minima. To directly smooth the global flatness, we propose the more significant FedGAM-CV, which employs control variables to correct local updates, guiding each client to train models in a globally flat direction. Experiments on three datasets (CIFAR-10, MNIST, and FashionMNIST) demonstrate that our proposed algorithms outperform existing FL baselines, effectively finding flat minima and addressing the client drift problem.
引用
收藏
页数:19
相关论文
共 50 条
  • [1] Acar DAE, 2021, Arxiv, DOI [arXiv:2111.04263, DOI 10.48550/ARXIV.2111.04263]
  • [2] The applications of machine learning techniques in medical data processing based on distributed computing and the Internet of Things
    Aminizadeh, Sarina
    Heidari, Arash
    Toumaj, Shiva
    Darbandi, Mehdi
    Navimipour, Nima Jafari
    Rezaei, Mahsa
    Talebi, Samira
    Azad, Poupak
    Unal, Mehmet
    [J]. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2023, 241
  • [3] Improving Generalization in Federated Learning by Seeking Flat Minima
    Caldarola, Debora
    Caputo, Barbara
    Ciccone, Marco
    [J]. COMPUTER VISION, ECCV 2022, PT XXIII, 2022, 13683 : 654 - 672
  • [4] AI in Finance: Challenges, Techniques, and Opportunities
    Cao, Longbing
    [J]. ACM COMPUTING SURVEYS, 2023, 55 (03)
  • [5] Understanding and Creating Art with AI: Review and Outlook
    Cetinic, Eva
    She, James
    [J]. ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2022, 18 (02)
  • [6] Entropy-SGD: biasing gradient descent into wide valleys
    Chaudhari, Pratik
    Choromanska, Anna
    Soatto, Stefano
    LeCun, Yann
    Baldassi, Carlo
    Borgs, Christian
    Chayes, Jennifer
    Sagun, Levent
    Zecchina, Riccardo
    [J]. JOURNAL OF STATISTICAL MECHANICS-THEORY AND EXPERIMENT, 2019, 2019 (12):
  • [7] Chen H.-Y, 2020, arXiv
  • [8] FedGAMMA: Federated Learning With Global Sharpness-Aware Minimization
    Dai, Rong
    Yang, Xun
    Sun, Yan
    Shen, Li
    Tian, Xinmei
    Wang, Meng
    Zhang, Yongdong
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (12) : 17479 - 17492
  • [9] Deng L., 2012, IEEE Signal Processing Magazine, V29, P141, DOI DOI 10.1109/MSP.2012.2211477
  • [10] Foret P., 2021, INT C LEARN REPR