Overcoming Noisy Labels and Non-IID Data in Edge Federated Learning

被引:5
作者
Xu, Yang [1 ,2 ]
Liao, Yunming [1 ,2 ]
Wang, Lun [1 ,2 ]
Xu, Hongli [1 ,2 ]
Jiang, Zhida [1 ,2 ]
Zhang, Wuyang [3 ]
机构
[1] Univ Sci & Technol China, Sch Comp Sci & Technol, Hefei 230027, Anhui, Peoples R China
[2] Univ Sci & Technol China, Suzhou Inst Adv Res, Suzhou 215123, Jiangsu, Peoples R China
[3] Meta Inc, Menlo Pk, CA 94025 USA
基金
美国国家科学基金会;
关键词
Noise measurement; Training; Data models; Computational modeling; Noise; Servers; Mobile computing; Federated learning; noisy labels; Non-IID data; edge computing;
D O I
10.1109/TMC.2024.3398801
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL) enables edge devices to cooperatively train models without exposing their raw data. However, implementing a practical FL system at the network edge mainly faces three challenges: label noise, data non-IIDness, and device heterogeneity, which seriously harm model performance and slow down convergence speed. Unfortunately, none of the existing works tackle all three challenges simultaneously. To this end, we develop a novel FL system, called Aorta, which features adaptive dataset construction and aggregation weight assignment. On each client, Aorta first calibrates potentially noisy labels and then constructs a training dataset with low noise, balanced distribution, and proper size. To fully utilize limited data on clients, we propose a global model guided method to select clean data and progressively correct noisy labels. To achieve balanced class distribution and proper dataset size, we propose a distribution-and-capability-aware data augmentation method to generate local training data. On the server, Aorta assigns aggregation weights based on the quality of local models to ensure that high-quality models have a greater influence on the global model. The model quality is measured through its cosine similarity with a benchmark model, which is trained on a clean and balanced dataset. We conduct extensive experiments on four datasets with various settings, including different noise types/ratios and non-IID types/levels. Compared to the baselines, Aorta improves model accuracy up to 9.8% on the datasets with moderate noise and non-IIDness, while providing a speedup of 4.2x on average when achieving the same target accuracy.
引用
收藏
页码:11406 / 11421
页数:16
相关论文
共 71 条
[1]  
Acar D. A. E., 2020, P INT C LEARN REPR, P1
[2]  
[Anonymous], 2006, SemiSupervised Learning
[3]  
Bai Y., 2021, PROC INT C NEURAL IN, p24 392
[4]  
Chen PF, 2021, AAAI CONF ARTIF INTE, V35, P11442
[5]   Randaugment: Practical automated data augmentation with a reduced search space [J].
Cubuk, Ekin D. ;
Zoph, Barret ;
Shlens, Jonathon ;
Le, Quoc, V .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, :3008-3017
[6]  
Cubuk ED, 2019, Arxiv, DOI arXiv:1805.09501
[7]   A Survey of Deep Learning and Its Applications: A New Paradigm to Machine Learning [J].
Dargan, Shaveta ;
Kumar, Munish ;
Ayyagari, Maruthi Rohit ;
Kumar, Gulshan .
ARCHIVES OF COMPUTATIONAL METHODS IN ENGINEERING, 2020, 27 (04) :1071-1092
[8]   Fed-DR-Filter: Using global data representation to reduce the impact of noisy labels on the performance of federated learning [J].
Duan, Shaoming ;
Liu, Chuanyi ;
Cao, Zhengsheng ;
Jin, Xiaopeng ;
Han, Peiyi .
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2022, 137 :336-348
[9]   Robust Heterogeneous Federated Learning under Data Corruption [J].
Fang, Xiuwen ;
Ye, Mang ;
Yang, Xiyuan .
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, :4997-5007
[10]   Robust Federated Learning with Noisy and Heterogeneous Clients [J].
Fang, Xiuwen ;
Ye, Mang .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :10062-10071