Towards Resource-Efficient Edge AI: From Federated Learning to Semi-Supervised Model Personalization

被引:4
作者
Zhang, Zhaofeng [1 ]
Yue, Sheng [1 ,2 ]
Zhang, Junshan [1 ,3 ]
机构
[1] Arizona State Univ, Sch Elect Comp & Energy Engn, Tempe, AZ 85281 USA
[2] Tsinghua Univ, Dept Comp Sci & Technol, Beijing 100190, Peoples R China
[3] Univ Calif Davis, Coll Engn Comp & Energy Engn, Davis, CA 95616 USA
基金
美国国家科学基金会;
关键词
Computational modeling; Data models; Training; Adaptation models; Servers; Performance evaluation; Internet of Things; Device heterogeneity; edge intelligence; federated learning; semi-supervised learning;
D O I
10.1109/TMC.2023.3316189
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A central question in edge intelligence is "how can an edge device learn its local model with limited data and constrained computing capacity?" In this study, we explore the approach where a global model initialization is first obtained by running federated learning (FL) across multiple edge devices, based on which a semi-supervised algorithm is devised for a single edge device to carry out quick adaptation with its local data. Specifically, to account for device heterogeneity and resource constraints, a global model is first trained via FL, where each device conducts multiple local updates only for its customized subnet. A subset of devices can be selected to upload updates for aggregation during each training round. Further, device scheduling is optimized to minimize the training loss of FL, subject to resource constraints, based on the carefully crafted reward function defined as the one-round progress of FL each device can provide. We examine the convergence behavior of FL for the general non-convex case. For semi-supervised model personalization, we use the FL-based model initialization as a teacher network to impute soft labels on unlabeled data, thereby addressing the insufficiency of labeled data. Experiments are conducted to evaluate the performance of the proposed algorithms.
引用
收藏
页码:6104 / 6115
页数:12
相关论文
共 50 条
[1]  
Chen T, 2020, Arxiv, DOI [arXiv:2006.10029, 10.48550/arXiv.2006.10029]
[2]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[3]  
Dun C, 2022, Arxiv, DOI arXiv:2107.00961
[4]  
Fallah A, 2020, Arxiv, DOI arXiv:2002.07948
[5]  
Finn C, 2017, PR MACH LEARN RES, V70
[6]  
Gong XW, 2020, IEEE INFOCOM SER, P2629, DOI [10.1109/infocom41043.2020.9155272, 10.1109/INFOCOM41043.2020.9155272]
[7]  
Grandvalet Y., 2004, P NEURIPS
[8]  
Hinton G, 2015, Arxiv, DOI arXiv:1503.02531
[9]  
Horváth S, 2021, ADV NEUR IN, V34
[10]   Structural learning with forgetting [J].
Ishikawa, M .
NEURAL NETWORKS, 1996, 9 (03) :509-521