FedSNN: Training Slimmable Neural Network With Federated Learning in Edge Computing

被引:0
作者
Xu, Yang [1 ,2 ]
Liao, Yunming [1 ,2 ]
Xu, Hongli [1 ,2 ]
Wang, Zhiyuan [1 ,2 ]
Wang, Lun [1 ,2 ]
Liu, Jianchun [1 ,2 ]
Qian, Chen [3 ]
机构
[1] Univ Sci & Technol China, Sch Comp Sci & Technol, Hefei 230027, Anhui, Peoples R China
[2] Univ Sci & Technol China, Suzhou Inst Adv Study, Suzhou 215123, Jiangsu, Peoples R China
[3] Univ Calif Santa Cruz, Jack Baskin Sch Engn, Dept Comp Sci & Engn, Santa Cruz, CA 95064 USA
来源
IEEE TRANSACTIONS ON NETWORKING | 2025年 / 33卷 / 01期
基金
美国国家科学基金会;
关键词
Edge computing; slimmable neural network; federated learning; data and system heterogeneity;
D O I
10.1109/TNET.2024.3487582
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
To provide a flexible tradeoff between inference accuracy and resource requirement at runtime, the slimmable neural network (SNN), a single network executable at different widths with the same deploying and management cost as that of a single model, has been proposed. However, how to effectively train SNN among massive devices in edge computing without revealing their local data remains an open problem. To this end, we leverage a novel distributed machine learning paradigm, i.e., federated learning, to realize effective on-device SNN training. As current FL schemes often train only one model with fixed architecture, and the existing SNN training algorithm is resource- intensive, integrating FL and SNN is non-trivial. Furthermore, two intrinsic features in edge computing, i.e., data and system heterogeneity, exacerbate the difficulty. Motivated by this, we redesign the model distribution, local training, and model aggregation phases in traditional FL, and propose FedSNN, a framework that ensures all widths in SNN can obtain high accuracy with less resource consumption. Specifically, for devices with heterogeneous training capacities and data distributions, the parameter server will distribute each of them with one proper width for adaptive local training guided by their uploaded model features, and their trained models will be weighted-averaged using the proposed multi-width SNN aggregation to improve their statistical utility. Extensive experiments on a distributed testbed show that FedSNN improves the model accuracy by about 2.18%-8.1%, and accelerates training by about 1.31x-6.84x, compared with existing solutions.
引用
收藏
页码:414 / 429
页数:16
相关论文
共 69 条
[1]   Mobile Edge Computing: A Survey [J].
Abbas, Nasir ;
Zhang, Yan ;
Taherkordi, Amir ;
Skeie, Tor .
IEEE INTERNET OF THINGS JOURNAL, 2018, 5 (01) :450-465
[2]  
Alam S, 2022, ADV NEUR IN
[3]  
Baek H., 2021, ARXIV
[4]   Joint Superposition Coding and Training for Federated Learning over Multi-Width Neural Networks [J].
Baek, Hankyul ;
Yun, Won Joon ;
Kwak, Yunseok ;
Jung, Soyi ;
Ji, Mingyue ;
Bennis, Mehdi ;
Park, Jihong ;
Kim, Joongheon .
IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2022), 2022, :1729-1738
[5]  
Bonawitz K., 2019, Proc Mach Learn Syst, DOI 10.48550/arXiv.1902.01046
[6]  
Bottou L., 2012, Stochastic Gradient Descent Tricks, P421, DOI DOI 10.1007/978-3-642-35289-8_25
[7]   "Learning-Compression" Algorithms for Neural Net Pruning [J].
Carreira-Perpinan, Miguel A. ;
Idelbayev, Yerlan .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8532-8541
[8]   Communication-efficient federated learning [J].
Chen, Mingzhe ;
Shlezinger, Nir ;
Poor, H. Vincent ;
Eldar, Yonina C. ;
Cui, Shuguang .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2021, 118 (17)
[9]   Joslim: Joint Widths and Weights Optimization for Slimmable Neural Networks [J].
Chin, Ting-Wu ;
Morcos, Ari S. ;
Marculescu, Diana .
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021: RESEARCH TRACK, PT III, 2021, 12977 :119-134
[10]  
Ciresan D, 2012, PROC CVPR IEEE, P3642, DOI 10.1109/CVPR.2012.6248110