Genetic CFL: Hyperparameter Optimization in Clustered Federated Learning

被引:43
作者
Agrawal, Shaashwat [1 ]
Sarkar, Sagnik [1 ]
Alazab, Mamoun [2 ]
Maddikunta, Praveen Kumar Reddy [3 ]
Gadekallu, Thippa Reddy [3 ]
Quoc-Viet Pham [4 ]
机构
[1] Vellore Inst Technol, Sch Comp Sci & Engn, Vellore 632014, Tamil Nadu, India
[2] Charles Darwin Univ, Coll Engn, IT & Environm, Casuarina, NT 0909, Australia
[3] Vellore Inst Technol, Sch Informat Technol & Engn, Vellore 632014, Tamil Nadu, India
[4] Pusan Natl Univ, Korean Southeast Ctr Ind Revolut Leader Educ 4, Busan 46241, South Korea
关键词
Deep learning;
D O I
10.1155/2021/7156420
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Federated learning (FL) is a distributed model for deep learning that integrates client-server architecture, edge computing, and real-time intelligence. FL has the capability of revolutionizing machine learning (ML) but lacks in the practicality of implementation due to technological limitations, communication overhead, non-IID (independent and identically distributed) data, and privacy concerns. Training a ML model over heterogeneous non-IID data highly degrades the convergence rate and performance. The existing traditional and clustered FL algorithms exhibit two main limitations, including inefficient client training and static hyperparameter utilization. To overcome these limitations, we propose a novel hybrid algorithm, namely, genetic clustered FL (Genetic CFL), that clusters edge devices based on the training hyperparameters and genetically modifies the parameters clusterwise. Then, we introduce an algorithm that drastically increases the individual cluster accuracy by integrating the density-based clustering and genetic hyperparameter optimization. The results are bench-marked using MNIST handwritten digit dataset and the CIFAR-10 dataset. The proposed genetic CFL shows significant improvements and works well with realistic cases of non-IID and ambiguous data. An accuracy of 99.79% is observed in the MNIST dataset and 76.88% in CIFAR-10 dataset with only 10 training rounds.
引用
收藏
页数:10
相关论文
共 32 条
[1]  
Alazab M., 2021, IEEE Trans. Ind. Inform.
[2]   Optimizing connection weights in neural networks using the whale optimization algorithm [J].
Aljarah, Ibrahim ;
Faris, Hossam ;
Mirjalili, Seyedali .
SOFT COMPUTING, 2018, 22 (01) :1-15
[3]  
Beruvides G, 2015, IEEE INTL CONF IND I, P1180, DOI 10.1109/INDIN.2015.7281903
[4]   Online detection of run out in microdrilling of tungsten and titanium alloys [J].
Beruvides, Gerardo ;
Quiza, Ramon ;
Rivas, Marcelino ;
Castao, Fernando ;
Haber, Rodolfo E. .
INTERNATIONAL JOURNAL OF ADVANCED MANUFACTURING TECHNOLOGY, 2014, 74 (9-12) :1567-1575
[5]   ST-DBSCAN: An algorithm for clustering spatial-temp oral data [J].
Birant, Derya ;
Kut, Alp .
DATA & KNOWLEDGE ENGINEERING, 2007, 60 (01) :208-221
[6]  
Bonawitz K, 2019, Arxiv, DOI arXiv:1902.01046
[7]  
Briggs C., 2020, IEEE IJCNN, P1, DOI DOI 10.1109/IJCNN48605.2020.9207469
[8]  
Briggs C, 2020, Arxiv, DOI [arXiv:2004.11791, DOI 10.48550/ARXIV.2004.11791]
[9]   Privacy-preserving and communication-efficient federated learning in Internet of Things [J].
Fang, Chen ;
Guo, Yuanbo ;
Hu, Yongjin ;
Ma, Bowen ;
Feng, Li ;
Yin, Anqi .
COMPUTERS & SECURITY, 2021, 103 (103)
[10]  
Ghosh A, 2021, Arxiv, DOI arXiv:2006.04088