FedKL: Tackling Data Heterogeneity in Federated Reinforcement Learning by Penalizing KL Divergence

被引:19
|
作者
Xie, Zhijie [1 ]
Song, Shenghui [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Dept Elect & Comp Engn, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Training; Convergence; Data models; Servers; Heuristic algorithms; Optimization; Linear programming; Federated reinforcement learning; data heterogeneity; policy gradient;
D O I
10.1109/JSAC.2023.3242734
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
One of the fundamental issues for Federated Learning (FL) is data heterogeneity, which causes accuracy degradation, slow convergence, and the communication bottleneck issue. Although the impact of data heterogeneity on supervised FL has been widely studied, the related investigation for Federated Reinforcement Learning (FRL) is still in its infancy. In this paper, we first define the type and level of data heterogeneity for FRL systems. By inspecting the connection between the global and local objective functions, we prove that local training can benefit the global objective, if the local update is properly penalized by the total variation (TV) distance between the local and global policies. A necessary condition for the global policy to be learn-able from the local environments is also derived, which is directly related to the heterogeneity level. Based on the theoretical result, a Kullback-Leibler (KL) divergence based penalty is proposed to directly constrain the model outputs in the distribution space and the convergence proof of the proposed algorithm is also provided. By jointly penalizing the divergence of the local policy from the global policy with a global penalty and penalizing each iteration of the local training with a local penalty, the proposed method achieves a better trade-off between training speed (step size) and convergence. Experiment results on two popular Reinforcement Learning (RL) experiment platforms demonstrate the advantage of the proposed algorithm over existing methods in accelerating and stabilizing the training process with heterogeneous data.
引用
收藏
页码:1227 / 1242
页数:16
相关论文
共 50 条
  • [21] Federated Learning with complete service commitment of data heterogeneity
    Zhou, Yizhi
    Wang, Junxiao
    Qin, Yuchen
    Kong, Xiangyu
    Xie, Xin
    Qi, Heng
    Zeng, Deze
    KNOWLEDGE-BASED SYSTEMS, 2025, 310
  • [22] Towards Taming the Resource and Data Heterogeneity in Federated Learning
    Chai, Zheng
    Fayyaz, Hannan
    Fayyaz, Zeshan
    Anwar, Ali
    Zhou, Yi
    Baracaldo, Nathalie
    Ludwig, Heiko
    Cheng, Yue
    PROCEEDINGS OF THE 2019 USENIX CONFERENCE ON OPERATIONAL MACHINE LEARNING, 2019, : 19 - 21
  • [23] Analysis and Optimization of Wireless Federated Learning With Data Heterogeneity
    Han, Xuefeng
    Li, Jun
    Chen, Wen
    Mei, Zhen
    Wei, Kang
    Ding, Ming
    Poor, H. Vincent
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2024, 23 (07) : 7728 - 7744
  • [24] Optimizing Federated Learning on Non-IID Data with Reinforcement Learning
    Wang, Hao
    Kaplan, Zakhary
    Niu, Di
    Li, Baochun
    IEEE INFOCOM 2020 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS, 2020, : 1698 - 1707
  • [25] Virtual Homogeneity Learning: Defending against Data Heterogeneity in Federated Learning
    Tang, Zhenheng
    Zhang, Yonggang
    Shi, Shaohuai
    He, Xin
    Han, Bo
    Chu, Xiaowen
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [26] Fed-RAC: Resource-Aware Clustering for Tackling Heterogeneity of Participants in Federated Learning
    Mishra, Rahul
    Gupta, Hari Prabhat
    Banga, Garvit
    Das, Sajal K.
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2024, 35 (07) : 1207 - 1220
  • [27] FLASH-RL: Federated Learning Addressing System and Static Heterogeneity using Reinforcement Learning
    Bouaziz, Sofiane
    Benmeziane, Hadjer
    Imine, Youcef
    Hamdad, Leila
    Niar, Smail
    Ouarnoughi, Hamza
    2023 IEEE 41ST INTERNATIONAL CONFERENCE ON COMPUTER DESIGN, ICCD, 2023, : 444 - 447
  • [28] Navigating Data Heterogeneity in Federated Learning: A Semi-Supervised Federated Object Detection
    Kim, Taehyeon
    Lin, Eric
    Lee, Junu
    Lau, Christian
    Mugunthan, Vaikkunth
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [29] Poster: Exploiting Data Heterogeneity for Performance and Reliability in Federated Learning
    Wang, Yuanli
    Kumar, Dhruv
    Chandra, Abhishek
    2020 IEEE/ACM SYMPOSIUM ON EDGE COMPUTING (SEC 2020), 2020, : 164 - 166
  • [30] Data Heterogeneity-Aware Personalized Federated Learning for Diagnosis
    Lin, Huiyan
    Li, Heng
    Jin, Haojin
    Yu, Xiangyang
    Yu, Kuai
    Liang, Chenhao
    Fu, Huazhu
    Liu, Jiang
    OPHTHALMIC MEDICAL IMAGE ANALYSIS, OMIA 2024, 2025, 15188 : 53 - 62