Enhancing Federated Learning Convergence With Dynamic Data Queue and Data-Entropy-Driven Participant Selection

被引:0
|
作者
Herath, Charuka [1 ]
Liu, Xiaolan [1 ]
Lambotharan, Sangarapillai [1 ]
Rahulamathavan, Yogachandran [1 ]
机构
[1] Loughborough Univ London, Inst Digital Technol, London E20 3BS, England
来源
IEEE INTERNET OF THINGS JOURNAL | 2025年 / 12卷 / 06期
基金
英国工程与自然科学研究理事会;
关键词
Data models; Convergence; Internet of Things; Distributed databases; Accuracy; Training; Mathematical models; Servers; Adaptation models; Data entropy; fairness FL; federated learning (FL); not identically and independently distributed (non-IID);
D O I
10.1109/JIOT.2024.3491034
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL) is a decentralized approach for collaborative model training on edge devices. This distributed method of model training offers advantages in privacy, security, regulatory compliance, and cost efficiency. Our emphasis in this research lies in addressing statistical complexity in FL, especially when the data stored locally across devices is not identically and independently distributed (non-IID). We have observed an accuracy reduction of up to approximately 10%-30%, particularly in skewed scenarios where each edge device trains with only 1 class of data. This reduction is attributed to weight divergence, quantified using the Euclidean distance between device-level class distributions and the population distribution, resulting in a bias term (delta(k)) . As a solution, we present a method to improve convergence in FL by creating a global subset of data on the server and dynamically distributing it across devices using a dynamic data queue-driven FL (DDFL). Next, we leverage Data Entropy metrics to observe the process during each training round and enable reasonable device selection for aggregation. Furthermore, we provide a convergence analysis of our proposed DDFL to justify their viability in practical FL scenarios, aiming for better device selection, a non-suboptimal global model, and faster convergence. We observe that our approach results in a substantial accuracy boost of approximately 5% for the MNIST dataset, around 18% for CIFAR-10, and 20% for CIFAR-100 with a 10% global subset of data, outperforming the state-of-the-art (SOTA) aggregation algorithms.
引用
收藏
页码:6646 / 6658
页数:13
相关论文
共 41 条
  • [1] FAST: Enhancing Federated Learning Through Adaptive Data Sampling and Local Training
    Wang, Zhiyuan
    Xu, Hongli
    Xu, Yang
    Jiang, Zhida
    Liu, Jianchun
    Chen, Suo
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2024, 35 (02) : 221 - 236
  • [2] Node Selection Toward Faster Convergence for Federated Learning on Non-IID Data
    Wu, Hongda
    Wang, Ping
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2022, 9 (05): : 3099 - 3111
  • [3] Participant Selection for Federated Learning With Heterogeneous Data in Intelligent Transport System
    Zhao, Jianxin
    Chang, Xinyu
    Feng, Yanhao
    Liu, Chi Harold
    Liu, Ningbo
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (01) : 1106 - 1115
  • [4] Data-Driven Participant Selection and Bandwidth Allocation for Heterogeneous Federated Edge Learning
    Albaseer, Abdullatif
    Abdallah, Mohamed
    Al-Fuqaha, Ala
    Erbad, Aiman
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2023, 53 (09): : 5848 - 5860
  • [5] DSFedCon: Dynamic Sparse Federated Contrastive Learning for Data-Driven Intelligent Systems
    Li, Zhengming
    Chen, Jiahui
    Zhang, Peifeng
    Huang, Huiwu
    Li, Guanbin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (02) : 3343 - 3355
  • [6] A Study of Enhancing Federated Learning on Non-IID Data with Server Learning
    Mai V.S.
    La R.J.
    Zhang T.
    IEEE Transactions on Artificial Intelligence, 2024, 5 (11): : 1 - 15
  • [7] Client Selection for Wireless Federated Learning With Data and Latency Heterogeneity
    Chen, Xiaobing
    Zhou, Xiangwei
    Zhang, Hongchao
    Sun, Mingxuan
    Vincent Poor, H.
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (19): : 32183 - 32196
  • [8] Client Selection for Federated Learning With Non-IID Data in Mobile Edge Computing
    Zhang, Wenyu
    Wang, Xiumin
    Zhou, Pan
    Wu, Weiwei
    Zhang, Xinglin
    IEEE ACCESS, 2021, 9 : 24462 - 24474
  • [9] Privacy-Preserving Data Selection for Horizontal and Vertical Federated Learning
    Zhang, Lan
    Li, Anran
    Peng, Hongyi
    Han, Feng
    Huang, Fan
    Li, Xiang-Yang
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2024, 35 (11) : 2054 - 2068
  • [10] Dynamic Adaptive Federated Learning on Local Long-Tailed Data
    Pu, Juncheng
    Fu, Xiaodong
    Dong, Hai
    Zhang, Pengcheng
    Liu, Li
    IEEE TRANSACTIONS ON SERVICES COMPUTING, 2024, 17 (06) : 3485 - 3498