Mutual Information Driven Federated Learning

被引:29
作者
Uddin, Md Palash [1 ]
Xiang, Yong [1 ]
Lu, Xuequan [1 ]
Yearwood, John [2 ]
Gao, Longxiang [1 ]
机构
[1] Deakin Univ, Deakin Blockchain Innovat Lab, Sch Informat Technol, Geelong, Vic 3220, Australia
[2] Deakin Univ, Sch Informat Technol, Geelong, Vic 3220, Australia
关键词
Data models; Training; Computational modeling; Servers; Mathematical model; Convergence; Analytical models; Distributed learning; federated learning; parallel optimization; data parallelism; information theory; mutual information; communication bottleneck; data heterogeneity; FEATURE-SELECTION;
D O I
10.1109/TPDS.2020.3040981
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Federated Learning (FL) is an emerging research field that yields a global trained model from different local clients without violating data privacy. Existing FL techniques often ignore the effective distinction between local models and the aggregated global model when doing the client-side weight update, as well as the distinction of local models for the server-side aggregation. In this article, we propose a novel FL approach with resorting to mutual information (MI). Specifically, in client-side, the weight update is reformulated through minimizing the MI between local and aggregated models and employing Negative Correlation Learning (NCL) strategy. In server-side, we select top effective models for aggregation based on the MI between an individual local model and its previous aggregated model. We also theoretically prove the convergence of our algorithm. Experiments conducted on MNIST, CIFAR-10, ImageNet, and the clinical MIMIC-III datasets manifest that our method outperforms the state-of-the-art techniques in terms of both communication and testing performance.
引用
收藏
页码:1526 / 1538
页数:13
相关论文
共 35 条
  • [11] Huang L, 2020, medRxiv, DOI DOI 10.1101/2020.03.05.20031898
  • [12] Jain P., 2017, P IARCS ANN C FDN SO
  • [13] Jain P, 2018, J MACH LEARN RES, V18
  • [14] MIMIC-III, a freely accessible critical care database
    Johnson, Alistair E. W.
    Pollard, Tom J.
    Shen, Lu
    Lehman, Li-wei H.
    Feng, Mengling
    Ghassemi, Mohammad
    Moody, Benjamin
    Szolovits, Peter
    Celi, Leo Anthony
    Mark, Roger G.
    [J]. SCIENTIFIC DATA, 2016, 3
  • [15] Johnson R., 2013, ADV NEURAL INFORM PR, P315, DOI DOI 10.5555/2999611.2999647
  • [16] Kone J., 2016, FEDERATED OPTIMIZATI, P1
  • [17] Efficient Mini-batch Training for Stochastic Optimization
    Li, Muu
    Zhang, Tong
    Chen, Yuqiang
    Smola, Alexander J.
    [J]. PROCEEDINGS OF THE 20TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING (KDD'14), 2014, : 661 - 670
  • [18] Accelerating Federated Learning via Momentum Gradient Descent
    Liu, Wei
    Chen, Li
    Chen, Yunfei
    Zhang, Wenyi
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2020, 31 (08) : 1754 - 1766
  • [19] Ensemble learning via negative correlation
    Liu, Y
    Yao, X
    [J]. NEURAL NETWORKS, 1999, 12 (10) : 1399 - 1404
  • [20] Simultaneous training of negatively correlated neural networks in an ensemble
    Liu, Y
    Yao, X
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 1999, 29 (06): : 716 - 725