Tackling the Non-IID Issue in Heterogeneous Federated Learning by Gradient Harmonization

被引:2
作者
Zhang, Xinyu [1 ]
Sun, Weiyu [1 ]
Chen, Ying [1 ]
机构
[1] Nanjing Univ, Sch Elect Sci & Engn, Nanjing 210023, Peoples R China
关键词
Training; Servers; Vectors; Optimization; Image classification; Social networking (online); Data models; Federated learning; non-IID issue; gradient conflict; gradient harmonization; robust server aggregation;
D O I
10.1109/LSP.2024.3430042
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Federated learning (FL) is a privacy-preserving paradigm for collaboratively training a global model from decentralized clients. However, the performance of FL is hindered by non-independent and identically distributed (non-IID) data and device heterogeneity. In this letter, we revisit this key challenge through the lens of gradient conflicts on the server side. Specifically, we first investigate the gradient conflict phenomenon among multiple clients and reveal that stronger heterogeneity leads to more severe gradient conflicts. To tackle this issue, we propose FedGH, a simple yet effective method that mitigates local drifts through Gradient Harmonization. This technique projects one gradient vector onto the orthogonal plane of the other within conflicting client pairs. Extensive experiments demonstrate that FedGH consistently enhances multiple state-of-the-art FL baselines across diverse benchmarks and non-IID scenarios. Moreover, FedGH yields more significant improvements in scenarios with stronger heterogeneity. As a plug-and-play module, FedGH can seamlessly integrate into any FL framework without requiring hyperparameter tuning.
引用
收藏
页码:2595 / 2599
页数:5
相关论文
共 36 条
  • [1] Deep Learning with Differential Privacy
    Abadi, Martin
    Chu, Andy
    Goodfellow, Ian
    McMahan, H. Brendan
    Mironov, Ilya
    Talwar, Kunal
    Zhang, Li
    [J]. CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, : 308 - 318
  • [2] Acar DAE, 2021, Arxiv, DOI [arXiv:2111.04263, DOI 10.48550/ARXIV.2111.04263]
  • [3] Improving Generalization in Federated Learning by Seeking Flat Minima
    Caldarola, Debora
    Caputo, Barbara
    Ciccone, Marco
    [J]. COMPUTER VISION, ECCV 2022, PT XXIII, 2022, 13683 : 654 - 672
  • [4] Caldas S, 2018, arXiv
  • [5] Cubuk ED, 2019, Arxiv, DOI [arXiv:1805.09501, DOI 10.48550/ARXIV.1805.09501]
  • [6] Dosovitskiy A, 2021, Arxiv, DOI [arXiv:2010.11929, 10.48550/arXiv.2010.11929]
  • [7] Geiping Jonas, 2020, Advances in neural information processing systems, V33
  • [8] Guo Yuhe, P MACHINE LEARNING R
  • [9] FedX: Unsupervised Federated Learning with Cross Knowledge Distillation
    Han, Sungwon
    Park, Sungwon
    Wu, Fangzhao
    Kim, Sundong
    Wu, Chuhan
    Xie, Xing
    Cha, Meeyoung
    [J]. COMPUTER VISION - ECCV 2022, PT XXX, 2022, 13690 : 691 - 707
  • [10] Hsu TMH, 2019, Arxiv, DOI arXiv:1909.06335