Multi-Granularity Weighted Federated Learning for Heterogeneous Edge Computing

被引:0
作者
Zhao, Yunfeng [1 ]
Qiu, Chao [1 ]
Cai, Shangxuan [1 ]
Liu, Zhicheng [1 ]
Wang, Yu [1 ]
Wang, Xiaofei [1 ]
Hu, Qinghua [1 ]
机构
[1] Tianjin Univ, Coll Intelligence & Comp, Tianjin 300072, Peoples R China
基金
中国国家自然科学基金; 美国国家科学基金会;
关键词
Data models; Computational modeling; Accuracy; Training; Analytical models; Smart phones; Federated learning; Collaboration; Visualization; Servers; Attention mechanism; edge computing; federated learning; model heterogeneity; multi-granularity learning; PRIVACY;
D O I
10.1109/TSC.2024.3495532
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL), an advanced variant of distributed machine learning, enables clients to collaboratively train a model without sharing raw data, thereby enhancing privacy, security, and reducing communication overhead. However, in edge computing scenarios, there is an increasing trend towards diversity, heterogeneity, and complexity in clients' data and models. The fundamental challenges, such as non-independent and identically distributed (non-IID) data and multi-granularity data accompanied by model heterogeneity, have become more evident and pose challenges to collaborative training among clients. In this paper, we refine the FL framework and propose the Multi-granularity Weighted Federated Learning (MGW-FL), emphasizing efficient collaborative training among clients with varied data granularities and diverse model scales across distinct data distributions. We introduce a distance-based FL mechanism designed for homogeneous clients, providing personalized models to mitigate the negative effects that non-IID data might have on model aggregation. Simultaneously, we propose an attention-weighted FL mechanism enhanced by a prior attention mechanism, facilitating knowledge transfer across clients with heterogeneous data granularities and model scales. Furthermore, we provide theoretical analyses of the convergence properties of the proposed MGW-FL method for both convex and non-convex models. Experimental results on five benchmark datasets demonstrate that, compared to baseline methods, MGW-FL significantly improves accuracy by almost 150% and convergence efficiency by nearly 20% on both IID and non-IID data.
引用
收藏
页码:270 / 287
页数:18
相关论文
共 54 条
  • [1] Aji Ablimit., 2010, CIKM, P629
  • [2] Arjovsky M, 2017, PR MACH LEARN RES, V70
  • [3] Bonawitz K, 2019, Arxiv, DOI arXiv:1902.01046
  • [4] Optimization Methods for Large-Scale Machine Learning
    Bottou, Leon
    Curtis, Frank E.
    Nocedal, Jorge
    [J]. SIAM REVIEW, 2018, 60 (02) : 223 - 311
  • [5] Large-Scale Machine Learning with Stochastic Gradient Descent
    Bottou, Leon
    [J]. COMPSTAT'2010: 19TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL STATISTICS, 2010, : 177 - 186
  • [6] An empirical comparison of combinations of evolutionary algorithms and neural networks for classification problems
    Cantú-Paz, E
    Kamath, C
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2005, 35 (05): : 915 - 927
  • [7] Communication-Efficient Federated Learning with Adaptive Parameter Freezing
    Chen, Chen
    Xu, Hong
    Wang, Wei
    Li, Baochun
    Li, Bo
    Chen, Li
    Zhang, Gong
    [J]. 2021 IEEE 41ST INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS 2021), 2021, : 1 - 11
  • [8] MetaFed: Federated Learning Among Federations With Cyclic Knowledge Distillation for Personalized Healthcare
    Chen, Yiqiang
    Lu, Wang
    Qin, Xin
    Wang, Jindong
    Xie, Xing
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, : 16671 - 16682
  • [9] Cheng P., 2021, P INT C LEARN REPR, P1
  • [10] Cohen G, 2017, IEEE IJCNN, P2921, DOI 10.1109/IJCNN.2017.7966217