Clustered federated learning based on nonconvex pairwise fusion

被引:3
作者
Yu, Xue [1 ]
Liu, Ziyi [1 ]
Wang, Wu [1 ]
Sun, Yifan [1 ,2 ]
机构
[1] Renmin Univ China, Ctr Appl Stat, Sch Stat, Beijing 100872, Peoples R China
[2] Beijing Adv Innovat Ctr Future Blockchain & Privac, Beijing, Peoples R China
关键词
Federated learning; Statistical heterogeneity; Pairwise fusion penalization; Efficient and robust; Optimization;
D O I
10.1016/j.ins.2024.120956
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This study investigates clustered federated learning (FL), a formulation of FL designed to handle non-i.i.d. data by partitioning devices into clusters, with each cluster optimizing a localized model for its data. We propose a clustered FL framework that incorporates a nonconvex penalty to pairwise differences of parameters. Without a priori knowledge of the set of devices in each cluster and the number of clusters, this framework can autonomously estimate cluster structures. To implement the proposed framework, we introduce a novel clustered FL method called Fusion Penalized Federated Clustering (FPFC). Building upon the standard alternating direction method of multipliers (ADMM), FPFC can perform partial updates at each communication round and allows parallel computation with a variable workload. These strategies significantly reduce the communication cost while ensuring privacy, thereby enhancing the practicality of FL. We also propose a new warmup strategy for hyperparameter tuning in FL settings and explore the asynchronous variant of FPFC (asyncFPFC). Theoretical analysis provides convergence guarantees for FPFC with general losses and establishes the statistical convergence rate under a linear model with squared loss. Extensive experiments have demonstrated the superiority of FPFC compared to current methods, including robustness and generalization capability.
引用
收藏
页数:35
相关论文
共 45 条
[1]   Distributed optimization and statistical learning via the alternating direction method of multipliers [J].
Boyd S. ;
Parikh N. ;
Chu E. ;
Peleato B. ;
Eckstein J. .
Foundations and Trends in Machine Learning, 2010, 3 (01) :1-122
[2]   COORDINATE DESCENT ALGORITHMS FOR NONCONVEX PENALIZED REGRESSION, WITH APPLICATIONS TO BIOLOGICAL FEATURE SELECTION [J].
Breheny, Patrick ;
Huang, Jian .
ANNALS OF APPLIED STATISTICS, 2011, 5 (01) :232-253
[3]   Communication-Efficient and Model-Heterogeneous Personalized Federated Learning via Clustered Knowledge Transfer [J].
Cho, Yae Jee ;
Wang, Jianyu ;
Chirvolu, Tarun ;
Joshi, Gauri .
IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2023, 17 (01) :234-247
[4]  
Dinh CT, 2020, ADV NEUR IN, V33
[5]  
DUA D., 2017, UCI machine learning repository
[6]  
Fallah A, 2020, ADV NEUR IN, V33
[7]   Variable selection via nonconcave penalized likelihood and its oracle properties [J].
Fan, JQ ;
Li, RZ .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2001, 96 (456) :1348-1360
[8]  
Feyzmahdavian HR, 2021, Arxiv, DOI arXiv:2109.04522
[9]   An Efficient Framework for Clustered Federated Learning [J].
Ghosh, Avishek ;
Chung, Jichan ;
Yin, Dong ;
Ramchandran, Kannan .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2022, 68 (12) :8076-8091
[10]   Network Lasso: Clustering and Optimization in Large Graphs [J].
Hallac, David ;
Leskovec, Jure ;
Boyd, Stephen .
KDD'15: PROCEEDINGS OF THE 21ST ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2015, :387-396