Optimizing Federated Learning on Non-IID Data with Reinforcement Learning

被引:0
|
作者
Wang, Hao [1 ]
Kaplan, Zakhary [1 ]
Niu, Di [2 ]
Li, Baochun [1 ]
机构
[1] Univ Toronto, Toronto, ON, Canada
[2] Univ Alberta, Edmonton, AB, Canada
来源
IEEE INFOCOM 2020 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS | 2020年
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
10.1109/infocom41043.2020.9155494
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The widespread deployment of machine learning applications in ubiquitous environments has sparked interests in exploiting the vast amount of data stored on mobile devices. To preserve data privacy, Federated Learning has been proposed to learn a shared model by performing distributed training locally on participating devices and aggregating the local models into a global one. However, due to the limited network connectivity of mobile devices, it is not practical for federated learning to perform model updates and aggregation on all participating devices in parallel. Besides, data samples across all devices are usually not independent and identically distributed (IID), posing additional challenges to the convergence and speed of federated learning. In this paper, we propose FAVOR, an experience-driven control framework that intelligently chooses the client devices to participate in each round of federated learning to counterbalance the bias introduced by non-IID data and to speed up convergence. Through both empirical and mathematical analysis, we observe an implicit connection between the distribution of training data on a device and the model weights trained based on those data, which enables us to profile the data distribution on that device based on its uploaded model weights. We then propose a mechanism based on deep Q-learning that learns to select a subset of devices in each communication round to maximize a reward that encourages the increase of validation accuracy and penalizes the use of more communication rounds. With extensive experiments performed in PyTorch, we show that the number of communication rounds required in federated learning can be reduced by up to 49% on the MNIST dataset, 23% on FashionMNIST, and 42% on CIFAR-10, as compared to the Federated Averaging algorithm.
引用
收藏
页码:1698 / 1707
页数:10
相关论文
共 50 条
  • [1] Federated learning on non-IID data: A survey
    Zhu, Hangyu
    Xu, Jinjin
    Liu, Shiqing
    Jin, Yaochu
    NEUROCOMPUTING, 2021, 465 : 371 - 390
  • [2] Adaptive Federated Learning With Non-IID Data
    Zeng, Yan
    Mu, Yuankai
    Yuan, Junfeng
    Teng, Siyuan
    Zhang, Jilin
    Wan, Jian
    Ren, Yongjian
    Zhang, Yunquan
    COMPUTER JOURNAL, 2023, 66 (11): : 2758 - 2772
  • [3] Federated Learning With Taskonomy for Non-IID Data
    Jamali-Rad, Hadi
    Abdizadeh, Mohammad
    Singh, Anuj
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (11) : 8719 - 8730
  • [4] Federated Learning With Non-IID Data: A Survey
    Lu, Zili
    Pan, Heng
    Dai, Yueyue
    Si, Xueming
    Zhang, Yan
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (11): : 19188 - 19209
  • [5] Non-IID Federated Learning
    Cao, Longbing
    IEEE INTELLIGENT SYSTEMS, 2022, 37 (02) : 14 - 15
  • [6] A Survey of Federated Learning on Non-IID Data
    HAN Xuming
    GAO Minghan
    WANG Limin
    HE Zaobo
    WANG Yanze
    ZTECommunications, 2022, 20 (03) : 17 - 26
  • [7] Differentially private federated learning with non-IID data
    Cheng, Shuyan
    Li, Peng
    Wang, Ruchuan
    Xu, He
    COMPUTING, 2024, 106 (07) : 2459 - 2488
  • [8] Fast converging Federated Learning with Non-IID Data
    Naas, Si -Ahmed
    Sigg, Stephan
    2023 IEEE 97TH VEHICULAR TECHNOLOGY CONFERENCE, VTC2023-SPRING, 2023,
  • [9] Adaptive Federated Deep Learning With Non-IID Data
    Zhang, Ze-Hui
    Li, Qing-Dan
    Fu, Yao
    He, Ning-Xin
    Gao, Tie-Gang
    Zidonghua Xuebao/Acta Automatica Sinica, 2023, 49 (12): : 2493 - 2506
  • [10] Federated Dictionary Learning from Non-IID Data
    Gkillas, Alexandros
    Ampeliotis, Dimitris
    Berberidis, Kostas
    2022 IEEE 14TH IMAGE, VIDEO, AND MULTIDIMENSIONAL SIGNAL PROCESSING WORKSHOP (IVMSP), 2022,