MHAT: An efficient model-heterogenous aggregation training scheme for federated learning

被引:108
作者
Hu, Li [1 ,2 ,3 ]
Yan, Hongyang [2 ]
Li, Lang [2 ]
Pan, Zijie [2 ]
Liu, Xiaozhang [1 ]
Zhang, Zulong [2 ]
机构
[1] Hainan Univ, Sch Comp Sci & Cyberspace Secur, Hannan, Peoples R China
[2] Guangzhou Univ, Sch Artificial Intelligence & Blockchain, Guangzhou, Peoples R China
[3] Peng Cheng Lab, Shenzhen, Peoples R China
基金
中国博士后科学基金;
关键词
Collaborative learning; Knowledge distillation; Heterogeneous; Communication;
D O I
10.1016/j.ins.2021.01.046
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated Learning allows multiple participants to jointly train a global model while guaranteeing the confidentiality and integrity of private datasets. However, current server aggregation algorithms for federated learning only focus on model parameters, resulting in heavy communication costs and low convergence speed. Most importantly, they are unable to handle the scenario wherein different clients hold different local models with various network architectures. In this paper, we view these challenges from an alternative perspective: we draw attention to what should be aggregated and how to improve convergence efficiency. Specifically, we propose MHAT, a novel model-heterogenous aggregation training federated learning scheme which exploits a technique of Knowledge Distillation (KD) to extract the update information of the heterogenous model of all clients and trains an auxiliary model on the server to realize information aggregation. MHAT relaxes clients from fixing on an unified model architecture and significantly reduces the required computing resources while maintaining acceptable model convergence accuracy. Various experiments verify the effectiveness and applicability of our proposed scheme. (c) 2021 Elsevier Inc. All rights reserved. Federated learning 11?3] is an emerging machine learning paradigm for decentralized data 14,5], which enables multiple parties to collaboratively train a global model without sharing their private data. In the canonical federated learning protocol 16], model parameter is the interactive information between clients and the server. At the beginning of each training round, the central server distributes current global model to all participants, and all participants update the local model using their private data. Then the server will collect updated model information from all parties and obtain the central global model via a weighted average aggregation algorithm. However, in a practical federated learning, several major challenges still remain unsolved when using model parameters
引用
收藏
页码:493 / 503
页数:11
相关论文
共 26 条
[21]  
Wang J, 2019, AAAI CONF ARTIF INTE, P1190
[22]   Federated Machine Learning: Concept and Applications [J].
Yang, Qiang ;
Liu, Yang ;
Chen, Tianjian ;
Tong, Yongxin .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2019, 10 (02)
[23]   A Gift from Knowledge Distillation: Fast Optimization, Network Minimization and Transfer Learning [J].
Yim, Junho ;
Joo, Donggyu ;
Bae, Jihoon ;
Kim, Junmo .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :7130-7138
[24]  
Yu T., 2020, arXiv
[25]   Deep Mutual Learning [J].
Zhang, Ying ;
Xiang, Tao ;
Hospedales, Timothy M. ;
Lu, Huchuan .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4320-4328
[26]  
Zhao Y., FEDERATED LEARNING N