Optimizing Federated Learning on Non-IID Data with Reinforcement Learning

被引:0
作者
Wang, Hao [1 ]
Kaplan, Zakhary [1 ]
Niu, Di [2 ]
Li, Baochun [1 ]
机构
[1] Univ Toronto, Toronto, ON, Canada
[2] Univ Alberta, Edmonton, AB, Canada
来源
IEEE INFOCOM 2020 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS | 2020年
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
10.1109/infocom41043.2020.9155494
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The widespread deployment of machine learning applications in ubiquitous environments has sparked interests in exploiting the vast amount of data stored on mobile devices. To preserve data privacy, Federated Learning has been proposed to learn a shared model by performing distributed training locally on participating devices and aggregating the local models into a global one. However, due to the limited network connectivity of mobile devices, it is not practical for federated learning to perform model updates and aggregation on all participating devices in parallel. Besides, data samples across all devices are usually not independent and identically distributed (IID), posing additional challenges to the convergence and speed of federated learning. In this paper, we propose FAVOR, an experience-driven control framework that intelligently chooses the client devices to participate in each round of federated learning to counterbalance the bias introduced by non-IID data and to speed up convergence. Through both empirical and mathematical analysis, we observe an implicit connection between the distribution of training data on a device and the model weights trained based on those data, which enables us to profile the data distribution on that device based on its uploaded model weights. We then propose a mechanism based on deep Q-learning that learns to select a subset of devices in each communication round to maximize a reward that encourages the increase of validation accuracy and penalizes the use of more communication rounds. With extensive experiments performed in PyTorch, we show that the number of communication rounds required in federated learning can be reduced by up to 49% on the MNIST dataset, 23% on FashionMNIST, and 42% on CIFAR-10, as compared to the Federated Averaging algorithm.
引用
收藏
页码:1698 / 1707
页数:10
相关论文
共 50 条
[21]   Data augmentation scheme for federated learning with non-IID data [J].
Tang L. ;
Wang D. ;
Liu S. .
Tongxin Xuebao/Journal on Communications, 2023, 44 (01) :164-176
[22]   A Study of Enhancing Federated Learning on Non-IID Data with Server Learning [J].
Mai V.S. ;
La R.J. ;
Zhang T. .
IEEE Transactions on Artificial Intelligence, 2024, 5 (11) :1-15
[23]   FedRL: Federated Learning with Non-IID Data via Review Learning [J].
Wang, Jinbo ;
Wang, Ruijin ;
Pei, Xikai .
2024 16TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND COMPUTING, ICMLC 2024, 2024, :115-120
[24]   Optimizing Federated Edge Learning on Non-IID Data via Neural Architecture Search [J].
Zhang, Feifei ;
Ge, Jidong ;
Wong, Chifong ;
Zhang, Sheng ;
Li, Chuanyi ;
Luo, Bin .
2021 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2021,
[25]   FedDRL: Deep Reinforcement Learning-based Adaptive Aggregation for Non-IID Data in Federated Learning [J].
Nang Hung Nguyen ;
Phi Le Nguyen ;
Duc Long Nguyen ;
Trung Thanh Nguyen ;
Thuy Dung Nguyen ;
Thanh Hung Nguyen ;
Huy Hieu Pham ;
Truong Thao Nguyen .
51ST INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, ICPP 2022, 2022,
[26]   A General Federated Learning Scheme with Blockchain on Non-IID Data [J].
Wu, Hao ;
Zhao, Shengnan ;
Zhao, Chuan ;
Jing, Shan .
INFORMATION SECURITY AND CRYPTOLOGY, INSCRYPT 2023, PT I, 2024, 14526 :126-140
[27]   Accelerating Federated learning on non-IID data against stragglers [J].
Zhang, Yupeng ;
Duan, Lingjie ;
Cheung, Ngai-Man .
2022 IEEE INTERNATIONAL CONFERENCE ON SENSING, COMMUNICATION, AND NETWORKING (SECON WORKSHOPS), 2022, :43-48
[28]   Inverse Distance Aggregation for Federated Learning with Non-IID Data [J].
Yeganeh, Yousef ;
Farshad, Azade ;
Navab, Nassir ;
Albarqouni, Shadi .
DOMAIN ADAPTATION AND REPRESENTATION TRANSFER, AND DISTRIBUTED AND COLLABORATIVE LEARNING, DART 2020, DCL 2020, 2020, 12444 :150-159
[29]   FEDERATED PAC-BAYESIAN LEARNING ON NON-IID DATA [J].
Zhao, Zihao ;
Liu, Yang ;
Ding, Wenbo ;
Zhang, Xiao-Ping .
2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, :5945-5949
[30]   FedProc: Prototypical contrastive federated learning on non-IID data [J].
Mu, Xutong ;
Shen, Yulong ;
Cheng, Ke ;
Geng, Xueli ;
Fu, Jiaxuan ;
Zhang, Tao ;
Zhang, Zhiwei .
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2023, 143 :93-104