FedKL: Tackling Data Heterogeneity in Federated Reinforcement Learning by Penalizing KL Divergence

被引:19
|
作者
Xie, Zhijie [1 ]
Song, Shenghui [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Dept Elect & Comp Engn, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Training; Convergence; Data models; Servers; Heuristic algorithms; Optimization; Linear programming; Federated reinforcement learning; data heterogeneity; policy gradient;
D O I
10.1109/JSAC.2023.3242734
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
One of the fundamental issues for Federated Learning (FL) is data heterogeneity, which causes accuracy degradation, slow convergence, and the communication bottleneck issue. Although the impact of data heterogeneity on supervised FL has been widely studied, the related investigation for Federated Reinforcement Learning (FRL) is still in its infancy. In this paper, we first define the type and level of data heterogeneity for FRL systems. By inspecting the connection between the global and local objective functions, we prove that local training can benefit the global objective, if the local update is properly penalized by the total variation (TV) distance between the local and global policies. A necessary condition for the global policy to be learn-able from the local environments is also derived, which is directly related to the heterogeneity level. Based on the theoretical result, a Kullback-Leibler (KL) divergence based penalty is proposed to directly constrain the model outputs in the distribution space and the convergence proof of the proposed algorithm is also provided. By jointly penalizing the divergence of the local policy from the global policy with a global penalty and penalizing each iteration of the local training with a local penalty, the proposed method achieves a better trade-off between training speed (step size) and convergence. Experiment results on two popular Reinforcement Learning (RL) experiment platforms demonstrate the advantage of the proposed algorithm over existing methods in accelerating and stabilizing the training process with heterogeneous data.
引用
收藏
页码:1227 / 1242
页数:16
相关论文
共 50 条
  • [1] Tackling Data Heterogeneity in Federated Learning with Class Prototypes
    Dai, Yutong
    Chen, Zeyuan
    Li, Junnan
    Heinecke, Shelby
    Sun, Lichao
    Xu, Ran
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 6, 2023, : 7314 - 7322
  • [2] An Aggregation-Free Federated Learning for Tackling Data Heterogeneity
    Wang, Yuan
    Fu, Huazhu
    Kanagavelu, Renuga
    Wei, Qingsong
    Liu, Yong
    Goh, Rick Siow Mong
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 26223 - 26232
  • [3] Tackling Data Heterogeneity in Federated Learning via Loss Decomposition
    Zeng, Shuang
    Guo, Pengxin
    Wang, Shuai
    Wang, Jianbo
    Zhou, Yuyin
    Qu, Liangqiong
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT X, 2024, 15010 : 707 - 717
  • [4] Rethinking Architecture Design for Tackling Data Heterogeneity in Federated Learning
    Qu, Liangqiong
    Zhou, Yuyin
    Liang, Paul Pu
    Xia, Yingda
    Wang, Feifei
    Adeli, Ehsan
    Li Fei-Fei
    Rubin, Daniel
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 10051 - 10061
  • [5] Federated Reinforcement Learning with Environment Heterogeneity
    Jin, Hao
    Peng, Yang
    Yang, Wenhao
    Wang, Shusen
    Zhang, Zhihua
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151 : 18 - 37
  • [6] Tackling data-heterogeneity variations in federated learning via adaptive aggregate weights
    Yin, Qiaoyun
    Feng, Zhiyong
    Li, Xiaohong
    Chen, Shizhan
    Wu, Hongyue
    Han, Gaoyong
    KNOWLEDGE-BASED SYSTEMS, 2024, 304
  • [7] FedClust: Tackling Data Heterogeneity in Federated Learning through Weight-Driven Client Clustering
    Islam, Md Sirajul
    Javaherian, Simin
    Xu, Fei
    Yuan, Xu
    Chen, Li
    Tzeng, Nian-Feng
    53RD INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, ICPP 2024, 2024, : 474 - 483
  • [8] Rethinking the Data Heterogeneity in Federated Learning
    Wang, Jiayi
    Wang, Shiqiang
    Chen, Rong-Rong
    Ji, Mingyue
    FIFTY-SEVENTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, IEEECONF, 2023, : 624 - 628
  • [9] Linear Convergence in Federated Learning: Tackling Client Heterogeneity and Sparse Gradients
    Mitra, Aritra
    Jaafar, Rayana
    Pappas, George J.
    Hassani, Hamed
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [10] Tackling heterogeneity in medical federated learning via aligning vision transformers
    Darzi, Erfan
    Shen, Yiqing
    Ou, Yangming
    van Ooijen, P. M. A.
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2024, 155