FedGAMMA: Federated Learning With Global Sharpness-Aware Minimization

被引:8
|
作者
Dai, Rong [1 ]
Yang, Xun [1 ]
Sun, Yan [2 ]
Shen, Li [3 ]
Tian, Xinmei [1 ]
Wang, Meng [4 ]
Zhang, Yongdong [1 ]
机构
[1] Univ Sci & Technol China, Sch Informat Sci & Technol, Hefei 230026, Peoples R China
[2] Univ Sydney, Sch Comp Sci, Sydney, NSW 2008, Australia
[3] JD Explore Acad, Beijing 100000, Peoples R China
[4] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Hefei 230009, Peoples R China
基金
中国国家自然科学基金;
关键词
Training; Servers; Federated learning; Data models; Minimization; Degradation; Convergence; Client-drift; deep learning; distributed learning; federated learning (FL);
D O I
10.1109/TNNLS.2023.3304453
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Federated learning (FL) is a promising framework for privacy-preserving and distributed training with decentralized clients. However, there exists a large divergence between the collected local updates and the expected global update, which is known as the client drift and mainly caused by heterogeneous data distribution among clients, multiple local training steps, and partial client participation training. Most existing works tackle this challenge based on the empirical risk minimization (ERM) rule, while less attention has been paid to the relationship between the global loss landscape and the generalization ability. In this work, we propose FedGAMMA, a novel FL algorithm with Global sharpness-Aware MiniMizAtion to seek a global flat landscape with high performance. Specifically, in contrast to FedSAM which only seeks the local flatness and still suffers from performance degradation when facing the client-drift issue, we adopt a local varieties control technique to better align each client's local updates to alleviate the client drift and make each client heading toward the global flatness together. Finally, extensive experiments demonstrate that FedGAMMA can substantially outperform several existing FL baselines on various datasets, and it can well address the client-drift issue and simultaneously seek a smoother and flatter global landscape.
引用
收藏
页码:17479 / 17492
页数:14
相关论文
共 50 条
  • [1] Federated Model-Agnostic Meta-Learning With Sharpness-Aware Minimization for Internet of Things Optimization
    Wu, Qingtao
    Zhang, Yong
    Liu, Muhua
    Zhu, Junlong
    Zheng, Ruijuan
    Zhang, Mingchuan
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (19): : 31317 - 31330
  • [2] Generalizable Prompt Learning via Gradient Constrained Sharpness-Aware Minimization
    Liu, Liangchen
    Wang, Nannan
    Zhou, Dawei
    Liu, Decheng
    Yang, Xi
    Gao, Xinbo
    Liu, Tongliang
    IEEE TRANSACTIONS ON MULTIMEDIA, 2025, 27 : 1100 - 1113
  • [3] Convergence of Sharpness-Aware Minimization with Momentum
    Pham Duy Khanh
    Luong, Hoang-Chau
    Mordukhovich, Boris S.
    Dat Ba Tran
    Truc Vo
    INFORMATION TECHNOLOGIES AND THEIR APPLICATIONS, PT II, ITTA 2024, 2025, 2226 : 123 - 132
  • [4] Sharpness-Aware Minimization and the Edge of Stability
    Long, Philip M.
    Bartlett, Peter L.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25 : 1 - 20
  • [5] Self-adaptive asynchronous federated optimizer with adversarial sharpness-aware minimization
    Zhang, Xiongtao
    Wang, Ji
    Bao, Weidong
    Xiao, Wenhua
    Zhang, Yaohong
    Liu, Lihua
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 161 : 638 - 654
  • [6] Adaptive Sharpness-Aware Minimization for Adversarial Domain Generalization
    Xie, Tianci
    Li, Tao
    Wu, Ruoxue
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024,
  • [7] SALR: Sharpness-Aware Learning Rate Scheduler for Improved Generalization
    Yue, Xubo
    Nouiehed, Maher
    Al Kontar, Raed
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (09) : 12518 - 12527
  • [8] Detection method of wheat rust based on transfer learning and sharpness-aware minimization
    Xu, Zhengguo
    Pan, Hengtuo
    Ye, Wei
    Xu, Zhuangwei
    Wang, Hongkai
    PLANT PATHOLOGY, 2023, 72 (02) : 353 - 360
  • [9] AdaSAM: Boosting sharpness-aware minimization with adaptive learning rate and momentum for neural networks
    Sun, Hao
    Shen, Li
    Zhong, Qihuang
    Ding, Liang
    Chen, Shixiang
    Sun, Jingwei
    Li, Jing
    Sun, Guangzhong
    Tao, Dacheng
    NEURAL NETWORKS, 2024, 169 : 506 - 519
  • [10] Sharpness-Aware Lookahead for Accelerating Convergence and Improving Generalization
    Tan, Chengli
    Zhang, Jiangshe
    Liu, Junmin
    Gong, Yihong
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 10375 - 10388