IFL-GAN: Improved Federated Learning Generative Adversarial Network With Maximum Mean Discrepancy Model Aggregation

被引:40
作者
Li, Wei [1 ,2 ]
Chen, Jinlin [3 ]
Wang, Zhenyu [4 ,5 ]
Shen, Zhidong [6 ]
Ma, Chao [6 ]
Cui, Xiaohui [6 ]
机构
[1] Jiangnan Univ, Sch Artificial Intelligence & Comp Sci, Sci Ctr Future Foods, Wuxi 214122, Jiangsu, Peoples R China
[2] Jiangnan Univ, Jiangsu Key Lab Media Design & Software Technol, Wuxi 214122, Jiangsu, Peoples R China
[3] Hong Kong Polytech Univ, Dept Comp, Hong Kong, Peoples R China
[4] Wuhan Univ, Sch Comp Sci, Wuhan 430072, Peoples R China
[5] Jiaxing Inst Future Food, Jiaxing 314005, Zhejiang, Peoples R China
[6] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan 430072, Peoples R China
关键词
Generative adversarial networks; Collaborative work; Data models; Training; Computational modeling; Distributed databases; Training data; Federated learning; generative adversarial network (GAN); maximum mean discrepancy (MMD); non-independent identically distributed (i; i; d; ) training data;
D O I
10.1109/TNNLS.2022.3167482
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The generative adversarial network (GAN) is usually built from the centralized, independent identically distributed (i.i.d.) training data to generate realistic-like instances. In real-world applications, however, the data may be distributed over multiple clients and hard to be gathered due to bandwidth, departmental coordination, or storage concerns. Although existing works, such as federated learning GAN (FL-GAN), adopt different distributed strategies to train GAN models, there are still limitations when data are distributed in a non-i.i.d. manner. These studies suffer from convergence difficulty, producing generated data with low quality. Fortunately, we found that these challenges are often due to the use of a federated averaging strategy to aggregate local GAN models' updates. In this article, we propose an alternative approach to tackling this problem, which learns a globally shared GAN model by aggregating locally trained generators' updates with maximum mean discrepancy (MMD). In this way, we term our approach improved FL-GAN (IFL-GAN). The MMD score helps each local GAN hold different weights, making the global GAN in IFL-GAN getting converged more rapidly than federated averaging. Extensive experiments on MNIST, CIFAR10, and SVHN datasets demonstrate the significant improvement of our IFL-GAN in both achieving the highest inception score and producing high-quality instances.
引用
收藏
页码:10502 / 10515
页数:14
相关论文
共 34 条
[1]  
[Anonymous], 2011, NEURIPS WORKSHOPS
[2]  
Barratt S., 2018, CoRR
[3]  
Bonawitz K., 2019, Proceedings of Machine Learning and Systems, V1, P374
[4]   THE COMPLEXITY OF COMPUTING A NASH EQUILIBRIUM [J].
Daskalakis, Constantinos ;
Goldberg, Paul W. ;
Papadimitriou, Christos H. .
SIAM JOURNAL ON COMPUTING, 2009, 39 (01) :195-259
[5]  
Goodfellow I, 2016, ADAPT COMPUT MACH LE, P1
[6]  
Goodfellow I.J., 2013, MULTIDIGIT NUMBER RE
[7]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672
[8]  
Gretton A, 2012, J MACH LEARN RES, V13, P723
[9]  
Hardy C., 2018, ARXIV181103850
[10]   Gossip Learning as a Decentralized Alternative to Federated Learning [J].
Hegedus, Istvan ;
Danner, Gabor ;
Jelasity, Mark .
DISTRIBUTED APPLICATIONS AND INTEROPERABLE SYSTEMS, DAIS 2019, 2019, 11534 :74-90