FedDT: A Communication-Efficient Federated Learning via Knowledge Distillation and Ternary Compression

被引:0
作者
He, Zixiao [1 ,2 ]
Zhu, Gengming [2 ]
Zhang, Shaobo [1 ,2 ]
Luo, Entao [3 ]
Zhao, Yijiang [2 ]
机构
[1] Hunan Univ Sci & Technol, Sanya Inst, Sanya 572024, Peoples R China
[2] Hunan Univ Sci & Technol, Sch Comp Sci & Engn, Xiangtan 411201, Peoples R China
[3] Hunan Univ Sci & Engn, Sch Elect & Informat Engn, Yongzhou 425199, Peoples R China
来源
ELECTRONICS | 2025年 / 14卷 / 11期
基金
中国国家自然科学基金;
关键词
federated learning; communication efficiency; data heterogeneity; knowledge distillation; ternary quantization;
D O I
10.3390/electronics14112183
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL) enables privacy-preserving collaborative training by iteratively aggregating locally trained model parameters on a central server while keeping raw data decentralized. However, FL faces critical challenges arising from data heterogeneity, model heterogeneity, and excessive communication costs. To address these issues, we propose a communication-efficient federated learning via knowledge distillation and ternary compression framework (FedDT). First, to mitigate the negative impact of data heterogeneity, we pre-train personalized heterogeneous teacher models for each client and employ knowledge distillation to transfer knowledge from teachers to student models, enhancing convergence speed and generalization capability. Second, to resolve model heterogeneity, we utilize the server-initialized global model as a shared student model across clients, where homogeneous student models mask local architectural variations to align feature representations. Finally, to reduce communication overhead, we introduce a two-level compression strategy that quantizes the distilled student model into ternary weight networks layer by layer, substantially decreasing parameter size. Comprehensive evaluations on both MNIST and Cifar10 datasets confirm that FedDT attains 7.85% higher model accuracy and reduces communication overhead by an average of 78% compared to baseline methods. This approach provides a lightweight solution for FL systems, significantly lowering communication costs while maintaining superior performance.
引用
收藏
页数:32
相关论文
共 49 条
[1]  
Bharati S., 2022, Int. J. Hybrid Intell. Syst, V18, P19, DOI [10.3233/HIS-220006, DOI 10.3233/HIS-220006]
[2]   Distributed Collaborative Learning with Representative Knowledge Sharing [J].
Casey, Joseph ;
Chen, Qianjiao ;
Fan, Mengchen ;
Geng, Baocheng ;
Shterenberg, Roman ;
Chen, Zhong ;
Li, Keren .
MATHEMATICS, 2025, 13 (06)
[3]  
Diao E., 2020, arXiv
[4]   Towards Personalized Federated Multi-Scenario Multi-Task Recommendation [J].
Ding, Yue ;
Ji, Yanbiao ;
Cai, Xun ;
Xin, Xin ;
Lu, Yuxiang ;
Huang, Suizhi ;
Liu, Chang ;
Gao, Xiaofeng ;
Murata, Tsuyoshi ;
Lu, Hongtao .
PROCEEDINGS OF THE EIGHTEENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, WSDM 2025, 2025, :429-438
[5]  
Hamer J., 2020, P MACHINE LEARNING R, P3973
[6]  
He C., 2020, Advances in neural information processing systems, V33, P14068
[7]  
Hinton G, 2015, Arxiv, DOI arXiv:1503.02531
[8]  
Huang YT, 2021, AAAI CONF ARTIF INTE, V35, P7865
[9]   Distillation-Based Semi-Supervised Federated Learning for Communication-Efficient Collaborative Training With Non-IID Private Data [J].
Itahara, Sohei ;
Nishio, Takayuki ;
Koda, Yusuke ;
Morikura, Masahiro ;
Yamamoto, Koji .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2023, 22 (01) :191-205
[10]  
Jeong E, 2023, Arxiv, DOI [arXiv:1811.11479, DOI 10.48550/ARXIV.1811.11479]