FLIGHT: Federated Learning with IRS for Grouped Heterogeneous Training

被引:0
作者
Yin T. [1 ]
Li L. [1 ]
Ma D. [1 ]
Lin W. [1 ]
Liang J. [1 ]
Han Z. [2 ,3 ]
机构
[1] School of Electronics and Information, Northwestern Polytechnical University, Xi’an
[2] Department of Electrical and Computer Engineering at the University of Houston, Houston, 77004, TX
[3] Department of Computer Science and Engineering, Kyung Hee University, Seoul
来源
Journal of Communications and Information Networks | 2022年 / 7卷 / 02期
基金
中国国家自然科学基金; 美国国家科学基金会;
关键词
decentralized aggrega-tion; federated learning; grouped learning; intelligent reflecting surfaces;
D O I
10.23919/jcin.2022.9815197
中图分类号
学科分类号
摘要
In recent years, federated learning (FL) has played an important role in private data-sensitive scenarios to perform learning tasks collectively without data exchange. However, due to the centralized model aggregation for heterogeneous devices in FL, the last updated model after local training delays the conver-gence, which increases the economic cost and dampens clients’ motivations for participating in FL. In addition, with the rapid development and application of intelligent reflecting surface (IRS) in the next-generation wireless communication, IRS has proven to be one effective way to enhance the communication quality. In this paper, we propose a framework of federated learning with IRS for grouped heterogeneous training (FLIGHT) to reduce the latency caused by the heterogeneous communication and computation of the clients. Specifically, we formulate a cost function and a greedy-based grouping strategy, which divides the clients into several groups to accelerate the convergence of the FL model. The simulation results verify the effectiveness of FLIGHT for accelerating the convergence of FL with heterogeneous clients. Besides the exemplified linear regression (LR) model and convolu-tional neural network (CNN), FLIGHT is also applicable to other learning models. © 2022, Posts and Telecom Press Co Ltd. All rights reserved.
引用
收藏
页码:135 / 146
页数:11
相关论文
共 50 条
[21]   Improving Federated Learning on Heterogeneous Data via Serial Pipeline Training and Global Knowledge Regularization [J].
Luo, Yiyang ;
Lu, Ting ;
Chang, Shan ;
Wang, Bingyue .
2022 IEEE 28TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS, ICPADS, 2022, :851-858
[22]   A novel staged training strategy leveraging knowledge distillation and model fusion for heterogeneous federated learning [J].
Wang, Debao ;
Guan, Shaopeng ;
Sun, Ruikang .
JOURNAL OF NETWORK AND COMPUTER APPLICATIONS, 2025, 236
[23]   Federated split GANs for collaborative training with heterogeneous devices [J].
Liang, Yilei ;
Kortoci, Pranvera ;
Zhou, Pengyuan ;
Lee, Lik-Hang ;
Mehrabi, Abbas ;
Hui, Pan ;
Tarkoma, Sasu ;
Crowcroft, Jon .
SOFTWARE IMPACTS, 2022, 14
[24]   The heterogeneous federated recommendation framework based on adversarial training [J].
Zhang, Suqi ;
Li, Yajuan ;
Zhao, Wang .
COMPLEX & INTELLIGENT SYSTEMS, 2025, 11 (09)
[25]   Empirical Analysis of Federated Learning in Heterogeneous Environments [J].
Abdelmoniem, Ahmed M. ;
Ho, Chen-Yu ;
Papageorgiou, Pantelis ;
Canini, Marco .
PROCEEDINGS OF THE 2022 2ND EUROPEAN WORKSHOP ON MACHINE LEARNING AND SYSTEMS (EUROMLSYS '22), 2022, :1-9
[26]   Heterogeneous Federated Learning Based on Graph Hypernetwork [J].
Xu, Zhengyi ;
Yang, Liu ;
Gu, Shiqiao .
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT III, 2023, 14256 :464-476
[27]   Federated learning with incremental clustering for heterogeneous data [J].
Espinoza Castellon, Fabiola ;
Mayoue, Aurelien ;
Sublemontier, Jacques-Henri ;
Gouy-Pailler, Cedric .
2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
[28]   Federated Deep Learning for Heterogeneous Edge Computing [J].
Ahmed, Khandaker Mamun ;
Imteaj, Ahmed ;
Amini, M. Hadi .
20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, :1146-1152
[29]   Exploiting features and logits in heterogeneous federated learning [J].
Chan, Yun-Hin ;
Ngai, Edith C. H. .
COMPUTER NETWORKS, 2025, 264
[30]   Distributional Knowledge Transfer for Heterogeneous Federated Learning [J].
Wang, Luau ;
Wang, Lijuan ;
Shcn, Jun .
2022 IEEE INTL CONF ON PARALLEL & DISTRIBUTED PROCESSING WITH APPLICATIONS, BIG DATA & CLOUD COMPUTING, SUSTAINABLE COMPUTING & COMMUNICATIONS, SOCIAL COMPUTING & NETWORKING, ISPA/BDCLOUD/SOCIALCOM/SUSTAINCOM, 2022, :747-754