Faster Convergence on Heterogeneous Federated Edge Learning: An Adaptive Clustered Data Sharing Approach

被引:0
作者
Hu, Gang [1 ]
Teng, Yinglei [1 ]
Wang, Nan [1 ]
Han, Zhu [2 ,3 ]
机构
[1] Beijing Univ Posts & Telecommun BUPT, Beijing Key Lab Work Safety Intelligent Monitoring, Beijing 100876, Peoples R China
[2] Univ Houston, Dept Elect & Comp Engn, Houston, TX 77004 USA
[3] Kyung Hee Univ, Dept Comp Sci & Engn, Seoul 446701, South Korea
基金
日本科学技术振兴机构; 中国国家自然科学基金; 国家重点研发计划;
关键词
Training; Data models; Accuracy; Optimization; Distributed databases; Convergence; Delays; Data privacy; Costs; Computational modeling; 6G; federated learning; non-IID data; multicasting; sidelink; data sharing;
D O I
10.1109/TMC.2025.3533566
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated Edge Learning (FEL) emerges as a pioneering distributed machine learning paradigm for the 6 G Hyper-Connectivity, harnessing data from the IoT devices while upholding data privacy. However, current FEL algorithms struggle with non-independent and non-identically distributed (non-IID) data, leading to elevated communication costs and compromised model accuracy. To address these statistical imbalances, we introduce a clustered data sharing framework, mitigating data heterogeneity by selectively sharing partial data from cluster heads to trusted associates through sidelink-aided multicasting. The collective communication pattern is integral to FEL training, where both cluster formation and the efficiency of communication and computation impact training latency and accuracy simultaneously. To tackle the strictly coupled data sharing and resource optimization, we decompose the optimization problem into the clients clustering and effective data sharing subproblems. Specifically, a distribution-based adaptive clustering algorithm (DACA) is devised basing on three deductive cluster forming conditions, which ensures the maximum sharing yield. Meanwhile, we design a stochastic optimization based joint computed frequency and shared data volume optimization (JFVO) algorithm, determining the optimal resource allocation with an uncertain objective function. The experiments show that the proposed framework facilitates FEL on non-IID datasets with faster convergence rate and higher model accuracy in a resource-limited environment.
引用
收藏
页码:5342 / 5356
页数:15
相关论文
共 48 条
[21]  
Li D., 2019, ADV NEURAL INF PROCE, P1
[22]   Federated Learning on Non-IID Data Silos: An Experimental Study [J].
Li, Qinbin ;
Diao, Yiqun ;
Chen, Quan ;
He, Bing Sheng .
2022 IEEE 38TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2022), 2022, :965-978
[23]  
Li T., 2020, Proc. Mach. Learn. Syst., V2, P429
[24]  
Li X., 2020, P INT C LEARN REPR A, P1
[25]   Social-Aware D2D Communications: Qualitative Insights and Quantitative Analysis [J].
Li, Yong ;
Wu, Ting ;
Hui, Pan ;
Jin, Depeng ;
Chen, Sheng .
IEEE COMMUNICATIONS MAGAZINE, 2014, 52 (06) :150-158
[26]   Federated Learning in Mobile Edge Networks: A Comprehensive Survey [J].
Lim, Wei Yang Bryan ;
Nguyen Cong Luong ;
Dinh Thai Hoang ;
Jiao, Yutao ;
Liang, Ying-Chang ;
Yang, Qiang ;
Niyato, Dusit ;
Miao, Chunyan .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2020, 22 (03) :2031-2063
[27]   Online Successive Convex Approximation for Two-Stage Stochastic Nonconvex Optimization [J].
Liu, An ;
Lau, Vincent K. N. ;
Zhao, Min-Jian .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2018, 66 (22) :5941-5955
[28]  
McMahan HB, 2017, PR MACH LEARN RES, V54, P1273
[29]  
Nocedal J, 2006, SPRINGER SER OPER RE, P1, DOI 10.1007/978-0-387-40065-5
[30]   Clustered Federated Learning: Model-Agnostic Distributed Multitask Optimization Under Privacy Constraints [J].
Sattler, Felix ;
Mueller, Klaus-Robert ;
Samek, Wojciech .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (08) :3710-3722