FedCD: Personalized Federated Learning via Collaborative Distillation

被引:6
作者
Ahmad, Sabtain [1 ]
Aral, Atakan [2 ]
机构
[1] Vienna Univ Technol, Inst Informat Syst Engn, Vienna, Austria
[2] Univ Vienna, Fac Comp Sci, Vienna, Austria
来源
2022 IEEE/ACM 15TH INTERNATIONAL CONFERENCE ON UTILITY AND CLOUD COMPUTING, UCC | 2022年
基金
奥地利科学基金会;
关键词
Federated learning; knowledge distillation; collaborative learning; model heterogeneity; personalized learning;
D O I
10.1109/UCC56403.2022.00036
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning enables the creation of a centralized global model by aggregating updates from the locally trained models from multiple clients. While powerful, such an architecture is limited to applications where the needs of heterogeneous clients can be served by a single global model. It does not cater to the scenarios where each client independently designs its own model. Task and data heterogeneity inherent to such scenarios demand each client to specialize in the local setting while still being able to collaborate and transfer the acquired knowledge to the rest of the federation without sharing the data or the model. In this work, we utilize ensemble and collaborative learning to design a framework that enables the training of personalized models for heterogeneous clients with different learning capacities using federated learning. Empirical evaluations performed on the CIFAR100 dataset demonstrate that our framework is able to consistently improve the performance of all the participating models and outperform the independently trained models on the complete training set without collaboration. We analyze that all participants benefit from collaborative distillation and boast an average 1.4% increase in performance. Moreover, a comparison with the state-of-the-art approaches demonstrates that our framework outperforms the Federated Learning and Federation Distillation methods by up to a 2x increase in the average global accuracy.
引用
收藏
页码:189 / 194
页数:6
相关论文
共 22 条
[1]  
Anil R., 2018, Large scale distributed neural network training through online distillation
[2]   Staleness Control for Edge Data Analytics [J].
Aral, Atakan ;
Erol-Kantarci, Melike ;
Brandic, Ivona .
PROCEEDINGS OF THE ACM ON MEASUREMENT AND ANALYSIS OF COMPUTING SYSTEMS, 2020, 4 (02)
[3]  
Guha N, 2019, WORKSHOP ML PHONE OT
[4]   Online Knowledge Distillation via Collaborative Learning [J].
Guo, Qiushan ;
Wang, Xinjiang ;
Wu, Yichao ;
Yu, Zhipeng ;
Liang, Ding ;
Hu, Xiaolin ;
Luo, Ping .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :11017-11026
[5]  
Hinton G., 2014, Distilling the knowledge in a neural network
[6]  
Hsu T.-M. H., 2019, arXiv
[7]  
Jeong E, 2018, WORKSHOP MACHINE LEA
[8]   Incentive Design for Efficient Federated Learning in Mobile Networks: A Contract Theory Approach [J].
Kang, Jiawen ;
Xiong, Zehui ;
Niyato, Dusit ;
Yu, Han ;
Liang, Ying-Chang ;
Kim, Dong In .
2019 IEEE VTS ASIA PACIFIC WIRELESS COMMUNICATIONS SYMPOSIUM (APWCS 2019), 2019,
[9]  
Karimireddy SP, 2020, PR MACH LEARN RES, V119
[10]   Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy [J].
Kuncheva, LI ;
Whitaker, CJ .
MACHINE LEARNING, 2003, 51 (02) :181-207