Performance Enhancement in Federated Learning by Reducing Class Imbalance of Non-IID Data

被引:12
作者
Seol, Mihye [1 ]
Kim, Taejoon [1 ]
机构
[1] Chungbuk Natl Univ, Sch Informat & Commun Engn, Chungju 28644, South Korea
基金
新加坡国家研究基金会;
关键词
federated learning; non-IID data; class imbalance mitigation; EDGE;
D O I
10.3390/s23031152
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Due to the distributed data collection and learning in federated learnings, many clients conduct local training with non-independent and identically distributed (non-IID) datasets. Accordingly, the training from these datasets results in severe performance degradation. We propose an efficient algorithm for enhancing the performance of federated learning by overcoming the negative effects of non-IID datasets. First, the intra-client class imbalance is reduced by rendering the class distribution of clients close to Uniform distribution. Second, the clients to participate in federated learning are selected to make their integrated class distribution close to Uniform distribution for the purpose of mitigating the inter-client class imbalance, which represents the class distribution difference among clients. In addition, the amount of local training data for the selected clients is finely adjusted. Finally, in order to increase the efficiency of federated learning, the batch size and the learning rate of local training for the selected clients are dynamically controlled reflecting the effective size of the local dataset for each client. In the performance evaluation on CIFAR-10 and MNIST datasets, the proposed algorithm achieves 20% higher accuracy than existing federated learning algorithms. Moreover, in achieving this huge accuracy improvement, the proposed algorithm uses less computation and communication resources compared to existing algorithms in terms of the amount of data used and the number of clients joined in the training.
引用
收藏
页数:16
相关论文
共 29 条
[1]   Finite-time analysis of the multiarmed bandit problem [J].
Auer, P ;
Cesa-Bianchi, N ;
Fischer, P .
MACHINE LEARNING, 2002, 47 (2-3) :235-256
[2]  
Briggs C., 2020, 2020 INT JOINT C NEU, DOI DOI 10.1109/IJCNN48605.2020.9207469
[3]  
Cisco, 2020, Cisco Annual Internet Report (2018-2023) White Paper
[4]   Astraea: Self-balancing Federated Learning for Improving Classification Accuracy of Mobile Deep Learning Applications [J].
Duan, Moming ;
Liu, Duo ;
Chen, Xianzhang ;
Tan, Yujuan ;
Ren, Jinting ;
Qiao, Lei ;
Liang, Liang .
2019 IEEE 37TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2019), 2019, :246-254
[5]  
Goyal P., 2017, ADV NEUR IN
[6]  
Hsu T.-M.H., 2019, arXiv
[7]  
Kopparapu K., 2020, ARXIV
[8]  
Krizhevsky A., 2009, Learning multiple layers of features from tiny images
[9]   ON INFORMATION AND SUFFICIENCY [J].
KULLBACK, S ;
LEIBLER, RA .
ANNALS OF MATHEMATICAL STATISTICS, 1951, 22 (01) :79-86
[10]  
Li Deng, 2012, IEEE Signal Processing Magazine, V29, P141, DOI [DOI 10.1109/MSP.2012.2211477, 10.1109/MSP.2012.2211477]