Two-stage model fusion scheme based on knowledge distillation for stragglers in federated learning

被引:0
作者
Xu, Jiuyun [1 ]
Li, Xiaowen [1 ]
Zhu, Kongshang [1 ]
Zhou, Liang [1 ]
Zhao, Yingzhi [1 ]
机构
[1] China Univ Petr East China, Qingdao Inst Software, Coll Comp Sci & Technol, 66 Changjiang West Rd, Qingdao 266580, Peoples R China
关键词
Federated learning; Straggler problem; Knowledge distillation; Heterogeneity; Training efficiency;
D O I
10.1007/s13042-024-02436-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Federated learning, as an emerging distributed learning paradigm, enables devices (also called clients) storing local data to collaboratively participate in a training task without the data leaving the devices, aiming to achieve the effect of integrating multiparty data while meeting privacy protection requirements. However, the participating clients are autonomous entities in a real-world environment, with heterogeneity and network instability, which leads to FL being plagued by stragglers when intermediate training results are synchronously interacted. To this end, this paper proposes a new FL scheme with a two-stage fusion process based on knowledge distillation, which transfers knowledge of straggler models to the global model without delaying the training speed, thus balancing efficiency and model performance. We have evaluated the proposed algorithm on three popular datasets. The experimental results show that FedTd improves training efficiency and maintains good model accuracy compared to baseline methods under heterogeneous conditions, exhibiting strong robustness against stragglers. By our approach, the running time can be accelerated by 1.97-3.32x\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3.32\times$$\end{document} under scenarios with higher level of data heterogeneity.
引用
收藏
页码:3067 / 3083
页数:17
相关论文
共 42 条
[1]   TiFL: A Tier-based Federated Learning System [J].
Chai, Zheng ;
Ali, Ahsan ;
Zawad, Syed ;
Treux, Stacey ;
Anwar, Ali ;
Barcaldo, Nathalie ;
Zhou, Yi ;
Ludwig, Heiko ;
Yan, Feng ;
Cheng, Yue .
PROCEEDINGS OF THE 29TH INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE PARALLEL AND DISTRIBUTED COMPUTING, HPDC 2020, 2020, :125-136
[2]   FedHe: Heterogeneous Models and Communication-Efficient Federated Learning [J].
Chan, Yun Hin ;
Ngai, Edith C. H. .
2021 17TH INTERNATIONAL CONFERENCE ON MOBILITY, SENSING AND NETWORKING (MSN 2021), 2021, :207-214
[3]   Asynchronous Online Federated Learning for Edge Devices with Non-IID Data [J].
Chen, Yujing ;
Ning, Yue ;
Slawski, Martin ;
Rangwala, Huzefa .
2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, :15-24
[4]  
Daliang L, 2019, ARXIV, DOI DOI 10.48550/ARXIV.1910.03581
[5]   Coded Federated Learning [J].
Dhakal, Sagar ;
Prakash, Saurav ;
Yona, Yair ;
Talwar, Shilpa ;
Himayat, Nageen .
2019 IEEE GLOBECOM WORKSHOPS (GC WKSHPS), 2019,
[6]  
Hahn S, 2019, ARXIV, DOI DOI 10.48550/ARXIV.1908.01851
[7]   Time Efficient Federated Learning with Semi-asynchronous Communication [J].
Hao, Jiangshan ;
Zhao, Yanchao ;
Zhang, Jiale .
2020 IEEE 26TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS), 2020, :156-163
[8]  
Hinton E. Geoffrey, 2015, COMPUT SCI, P1, DOI DOI 10.48550/ARXIV.1503.02531
[9]  
Hsu TMH, 2019, ARXIV, DOI DOI 10.48550/ARXIV.1909.06335
[10]   Refine Myself by Teaching Myself : Feature Refinement via Self-Knowledge Distillation [J].
Ji, Mingi ;
Shin, Seungjae ;
Hwang, Seunghyun ;
Park, Gibeom ;
Moon, Il-Chul .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :10659-10668