Enhancing Federated Learning with In-Cloud Unlabeled Data

被引:11
作者
Wang, Lun [1 ,3 ]
Xu, Yang [1 ,3 ]
Xu, Hongli [1 ,3 ]
Liu, Jianchun [2 ,3 ]
Wang, Zhiyuan [1 ,3 ]
Huang, Liusheng [1 ,3 ]
机构
[1] Univ Sci & Technol China, Sch Comp Sci & Technol, Hefei, Anhui, Peoples R China
[2] Univ Sci & Technol China, Sch Data Sci, Hefei, Anhui, Peoples R China
[3] Univ Sci & Technol China, Suzhou Inst Adv Res, Hefei, Anhui, Peoples R China
来源
2022 IEEE 38TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2022) | 2022年
基金
美国国家科学基金会;
关键词
Edge Computing; Federated Learning; Semi-supervised Learning; Pseudo-labeling; BIG DATA;
D O I
10.1109/ICDE53745.2022.00015
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Federated learning (FL) has been widely applied to collaboratively train deep learning (DL) models on massive end devices (i.e., clients). Due to the limited storage capacity and high labeling cost, there are always insufficient data stored and annotated on each client. Conversely, in cloud datacenters, there exist large-scale unlabeled data, which are easy to collect from public access (e.g., social media). Herein, upon the federated semi-supervised learning (FSSL) technology, we propose the Ada-FedSemi system, which leverages both on-device labeled data and in-cloud unlabeled data to boost the performance of DL models. Given the limited communication and massive quantity of the clients, in each training round, we decide to select partial clients to participate in FL, and their local models are aggregated by the parameter server (PS) to produce pseudo-labels for the unlabeled data, which are utilized to enhance the global model. Considering that the number of participating clients and the quality of pseudo-labels will have a significant impact on the training performance (e.g., efficiency and accuracy), we introduce a multi-armed bandit (MAB) based online algorithm to adaptively determine the participating fraction and confidence threshold during federated model training. Extensive experiments on benchmark models and datasets show that, given the same resource budget, the model trained by Ada-FedSemi achieves 3%14.8% higher test accuracy than that of the baseline methods. Besides, when achieving the same test accuracy, Ada-FedSemi saves up to 48% training cost, compared with the baselines.
引用
收藏
页码:136 / 149
页数:14
相关论文
共 55 条
[1]  
Alex K., 2009, LEARNING MULTIPLE LA
[2]  
Arazo E., 2020, IEEE IJCNN, P1, DOI 10.1109/IJCNN48605.2020.9207304
[3]  
Berthelot D, 2019, ADV NEUR IN, V32
[4]  
Besbes O, 2014, ADV NEUR IN, V27
[5]  
Bistritz I., 2020, Advances in Neural Information Processing Systems, V33
[6]  
Cascante-Bonilla P, 2021, AAAI CONF ARTIF INTE, V35, P6912
[7]  
Cesa-Bianchi Nicolo, 2017, Advances in Neural Information Processing Systems, V30
[8]  
Dettmers T., 2016, ICLR
[9]  
Erlingsson Ulfar, 2016, ARXIV161005755
[10]   A survey on deep learning techniques for image and video semantic segmentation [J].
Garcia-Garcia, Alberto ;
Orts-Escolano, Sergio ;
Oprea, Sergiu ;
Villena-Martinez, Victor ;
Martinez-Gonzalez, Pablo ;
Garcia-Rodriguez, Jose .
APPLIED SOFT COMPUTING, 2018, 70 :41-65