FedSwarm: An Adaptive Federated Learning Framework for Scalable AIoT

被引:1
作者
Du, Haizhou [1 ]
Ni, Chengdong [1 ]
Cheng, Chaoqian [1 ]
Xiang, Qiao [2 ]
Chen, Xi [3 ]
Liu, Xue [3 ]
机构
[1] Shanghai Univ Elect Power, Sch Comp Sci & Technol, Shanghai 201399, Peoples R China
[2] Xiamen Univ, Sch Informat, Xiamen 361005, Peoples R China
[3] McGill Univ, Sch Comp Sci, Montreal, PQ H3A 0E9, Canada
基金
国家重点研发计划;
关键词
Servers; Scalability; Adaptation models; Computer architecture; Internet of Things; Performance evaluation; Artificial intelligence; Artificial Intelligence of Things (AIoT); dynamic per-device server selection; federated learning (FL); local update adaptation; RESOURCE-ALLOCATION; TOPOLOGY;
D O I
10.1109/JIOT.2023.3321325
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL) is a key solution for datadriven the Artificial Intelligence of Things (AIoT). Although much progress has been made, scalability remains a core challenge for real-world FL deployments. Existing solutions either suffer from accuracy loss or do not fully address the connectivity dynamicity of FL systems. In this article, we tackle the scalability issue with a novel, adaptive FL framework called FedSwarm, which improves system scalability for AIoT by deploying multiple collaborative edge servers. FedSwarm has two novel features: 1) adaptiveness on the number of local updates and 2) dynamicity of the synchronization between edge devices and edge servers. We formulate FedSwarm as a local update adaptation and perdevice dynamic server selection problem and prove FedSwarm's convergence bound. We further design a control mechanism consisting of a learning-based algorithm for collaboratively providing local update adaptation on the servers' side and a bonus-based strategy for spurring dynamic per-device server selection on the devices' side. Our extensive evaluation shows that FedSwarm significantly outperforms other studies with better scalability, lower energy consumption, and higher model accuracy.
引用
收藏
页码:8268 / 8287
页数:20
相关论文
共 77 条
[1]  
[Anonymous], 2009, LEARNING MULTIPLE LA
[2]  
Badia AP, 2020, PR MACH LEARN RES, V119
[3]  
Beutel DJ, 2022, Arxiv, DOI arXiv:2007.14390
[4]  
Bonawitz K., 2019, Proc. Mach. Learn. Res, V1, P374, DOI 10.48550/arXiv.1902.01046
[5]  
Burda Yuri, 2018, P INT C LEARN REPR
[6]   Entropy-SGD: biasing gradient descent into wide valleys [J].
Chaudhari, Pratik ;
Choromanska, Anna ;
Soatto, Stefano ;
LeCun, Yann ;
Baldassi, Carlo ;
Borgs, Christian ;
Chayes, Jennifer ;
Sagun, Levent ;
Zecchina, Riccardo .
JOURNAL OF STATISTICAL MECHANICS-THEORY AND EXPERIMENT, 2019, 2019 (12)
[7]   On the Convergence of Decentralized Federated Learning Under Imperfect Information Sharing [J].
Chellapandi, Vishnu Pandi ;
Upadhyay, Antesh ;
Hashemi, Abolfazl ;
Zak, Stanislaw H. .
IEEE CONTROL SYSTEMS LETTERS, 2023, 7 :2982-2987
[8]  
Chellapandi VP, 2023, Arxiv, DOI arXiv:2303.10677
[9]   Synchronize Only the Immature Parameters: Communication-Efficient Federated Learning By Freezing Parameters Adaptively [J].
Chen, Chen ;
Xu, Hong ;
Wang, Wei ;
Li, Baochun ;
Li, Bo ;
Chen, Li ;
Zhang, Gong .
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2024, 35 (07) :1155-1173
[10]  
Chilimbi Trishul, 2014, Proceedings of the 11th USENIX Symposium on Operating Systems Design and Implementation (OSDI '14). OSDI '14, P571