FedDT: A Communication-Efficient Federated Learning via Knowledge Distillation and Ternary Compression

被引:0
作者
He, Zixiao [1 ,2 ]
Zhu, Gengming [2 ]
Zhang, Shaobo [1 ,2 ]
Luo, Entao [3 ]
Zhao, Yijiang [2 ]
机构
[1] Hunan Univ Sci & Technol, Sanya Inst, Sanya 572024, Peoples R China
[2] Hunan Univ Sci & Technol, Sch Comp Sci & Engn, Xiangtan 411201, Peoples R China
[3] Hunan Univ Sci & Engn, Sch Elect & Informat Engn, Yongzhou 425199, Peoples R China
基金
中国国家自然科学基金;
关键词
federated learning; communication efficiency; data heterogeneity; knowledge distillation; ternary quantization;
D O I
10.3390/electronics14112183
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL) enables privacy-preserving collaborative training by iteratively aggregating locally trained model parameters on a central server while keeping raw data decentralized. However, FL faces critical challenges arising from data heterogeneity, model heterogeneity, and excessive communication costs. To address these issues, we propose a communication-efficient federated learning via knowledge distillation and ternary compression framework (FedDT). First, to mitigate the negative impact of data heterogeneity, we pre-train personalized heterogeneous teacher models for each client and employ knowledge distillation to transfer knowledge from teachers to student models, enhancing convergence speed and generalization capability. Second, to resolve model heterogeneity, we utilize the server-initialized global model as a shared student model across clients, where homogeneous student models mask local architectural variations to align feature representations. Finally, to reduce communication overhead, we introduce a two-level compression strategy that quantizes the distilled student model into ternary weight networks layer by layer, substantially decreasing parameter size. Comprehensive evaluations on both MNIST and Cifar10 datasets confirm that FedDT attains 7.85% higher model accuracy and reduces communication overhead by an average of 78% compared to baseline methods. This approach provides a lightweight solution for FL systems, significantly lowering communication costs while maintaining superior performance.
引用
收藏
页数:32
相关论文
共 49 条
[31]  
Vargaftik S, 2022, PR MACH LEARN RES
[32]   HetVis: A Visual Analysis Approach for Identifying Data Heterogeneity in Horizontal Federated Learning [J].
Wang X. ;
Chen W. ;
Xia J. ;
Wen Z. ;
Zhu R. ;
Schreck T. .
IEEE Transactions on Visualization and Computer Graphics, 2023, 29 (01) :310-319
[33]   Communication-efficient federated learning via knowledge distillation [J].
Wu, Chuhan ;
Wu, Fangzhao ;
Lyu, Lingjuan ;
Huang, Yongfeng ;
Xie, Xing .
NATURE COMMUNICATIONS, 2022, 13 (01)
[34]   Federated Learning With Client Selection and Gradient Compression in Heterogeneous Edge Systems [J].
Xu, Yang ;
Jiang, Zhida ;
Xu, Hongli ;
Wang, Zhiyuan ;
Qian, Chen ;
Qiao, Chunming .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (05) :5446-5461
[35]   Federated Machine Learning: Concept and Applications [J].
Yang, Qiang ;
Liu, Yang ;
Chen, Tianjian ;
Tong, Yongxin .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2019, 10 (02)
[36]  
Yang Zhengyi, 2023, Advances in Neural
[37]  
Ye M., 2025, ACM Comput. Surv, V57, P1, DOI [10.1145/3720539, DOI 10.1145/3720539]
[38]  
Yue K, 2023, PROCEEDINGS OF THE 32ND USENIX SECURITY SYMPOSIUM, P6381
[39]   A survey on federated learning [J].
Zhang, Chen ;
Xie, Yu ;
Bai, Hang ;
Yu, Bin ;
Li, Weihong ;
Gao, Yuan .
KNOWLEDGE-BASED SYSTEMS, 2021, 216
[40]  
Zhang J, 2023, PROCEEDINGS OF THE 20TH USENIX SYMPOSIUM ON NETWORKED SYSTEMS DESIGN AND IMPLEMENTATION, NSDI 2023, P1057