A Decentralized Federated Learning Based on Node Selection and Knowledge Distillation

被引:6
作者
Zhou, Zhongchang [1 ]
Sun, Fenggang [1 ]
Chen, Xiangyu [1 ]
Zhang, Dongxu [2 ]
Han, Tianzhen [3 ]
Lan, Peng [1 ]
机构
[1] Shandong Agr Univ, Coll Informat Sci & Engn, Tai An 271018, Peoples R China
[2] Taishan Intelligent Mfg Ind Res Inst, Tai An 271000, Peoples R China
[3] Taian Chinamobile, Network Dept Optimizat Ctr, Tai An 271000, Peoples R China
关键词
federated learning; node selection; decentralized learning; knowledge distillation;
D O I
10.3390/math11143162
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Federated learning has become increasingly important for modern machine learning, especially for data privacy sensitive scenarios. Existing federated learning mainly adopts a central server-based network topology, however, the training process of which is susceptible to the central node. To address this problem, this article proposed a decentralized federated learning method based on node selection and knowledge distillation. Specifically, the central node in this method is variable, and it is selected by the indicator interaction between nodes. Meanwhile, the knowledge distillation mechanism is added to make the student model as close as possible to the teacher's network and ensure the model's accuracy. The experiments were conducted on the public MNIST, CIFAR-10, and FEMNIST datasets for both the Independent Identically Distribution (IID) setting and the non-IID setting. Numerical results show that the proposed method can achieve an improved accuracy as compared to the centralized federated learning method, and the computing time is reduced greatly with less accuracy loss as compared to the blockchain decentralized federated learning. Therefore, the proposed method guarantees the model effect while meeting the individual model requirements of each node and reducing the running time.
引用
收藏
页数:15
相关论文
共 28 条
[1]   Neural Fairness Blockchain Protocol Using an Elliptic Curves Lottery [J].
Caldarola, Fabio ;
D'Atri, Gianfranco ;
Zanardo, Enrico .
MATHEMATICS, 2022, 10 (17)
[2]  
Caldas S., 2018, arXiv
[3]   Toward Federated Learning With Byzantine and Inactive Users: A Game Theory Approach [J].
Chen, Xiangyu ;
Lan, Peng ;
Zhou, Zhongchang ;
Zhao, Angran ;
Zhou, Pengfei ;
Sun, Fenggang .
IEEE ACCESS, 2023, 11 :34138-34149
[4]  
He CY, 2020, Arxiv, DOI arXiv:1910.04956
[5]   An Efficiency-Boosting Client Selection Scheme for Federated Learning With Fairness Guarantee [J].
Huang, Tiansheng ;
Lin, Weiwei ;
Wu, Wentai ;
He, Ligang ;
Li, Keqin ;
Zomaya, Albert .
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2021, 32 (07) :1552-1564
[6]  
Jeong E, 2023, Arxiv, DOI arXiv:1811.11479
[7]   Advances and Open Problems in Federated Learning [J].
Kairouz, Peter ;
McMahan, H. Brendan ;
Avent, Brendan ;
Bellet, Aurelien ;
Bennis, Mehdi ;
Bhagoji, Arjun Nitin ;
Bonawitz, Kallista ;
Charles, Zachary ;
Cormode, Graham ;
Cummings, Rachel ;
D'Oliveira, Rafael G. L. ;
Eichner, Hubert ;
El Rouayheb, Salim ;
Evans, David ;
Gardner, Josh ;
Garrett, Zachary ;
Gascon, Adria ;
Ghazi, Badih ;
Gibbons, Phillip B. ;
Gruteser, Marco ;
Harchaoui, Zaid ;
He, Chaoyang ;
He, Lie ;
Huo, Zhouyuan ;
Hutchinson, Ben ;
Hsu, Justin ;
Jaggi, Martin ;
Javidi, Tara ;
Joshi, Gauri ;
Khodak, Mikhail ;
Konecny, Jakub ;
Korolova, Aleksandra ;
Koushanfar, Farinaz ;
Koyejo, Sanmi ;
Lepoint, Tancrede ;
Liu, Yang ;
Mittal, Prateek ;
Mohri, Mehryar ;
Nock, Richard ;
Ozgur, Ayfer ;
Pagh, Rasmus ;
Qi, Hang ;
Ramage, Daniel ;
Raskar, Ramesh ;
Raykova, Mariana ;
Song, Dawn ;
Song, Weikang ;
Stich, Sebastian U. ;
Sun, Ziteng ;
Suresh, Ananda Theertha .
FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2021, 14 (1-2) :1-210
[8]  
Krizhevsky A., 2009, LEARNING MULTIPLE LA
[9]   Gradient-based learning applied to document recognition [J].
Lecun, Y ;
Bottou, L ;
Bengio, Y ;
Haffner, P .
PROCEEDINGS OF THE IEEE, 1998, 86 (11) :2278-2324
[10]  
Li T., 2020, P MACHINE LEARNING S, V2, P429