An EMD-Based Adaptive Client Selection Algorithm for Federated Learning in Heterogeneous Data Scenarios

被引:6
作者
Chen, Aiguo [1 ]
Fu, Yang [1 ]
Sha, Zexin [1 ]
Lu, Guoming [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu, Peoples R China
来源
FRONTIERS IN PLANT SCIENCE | 2022年 / 13卷
基金
中国国家自然科学基金;
关键词
distributed conjoint analysis; federated learning; adaptive client selection; statistical heterogeneity; machine learning;
D O I
10.3389/fpls.2022.908814
中图分类号
Q94 [植物学];
学科分类号
071001 ;
摘要
Federated learning is a distributed machine learning framework that enables distributed nodes with computation and storage capabilities to train a global model while keeping distributed-stored data locally. This process can promote the efficiency of modeling while preserving data privacy. Therefore, federated learning can be widely applied in distributed conjoint analysis scenarios, such as smart plant protection systems, in which widely networked IoT devices are used to monitor the critical data of plant production to improve crop production. However, the data collected by different IoT devices can be dependent and identically distributed (non-IID), causing the challenge of statistical heterogeneity. Studies have also shown that statistical heterogeneity can lead to a marked decline in the efficiency of federated learning, making it challenging to apply in practice. To promote the efficiency of federated learning in statistical heterogeneity scenarios, an adaptive client selection algorithm for federated learning in statistical heterogeneous scenarios called ACSFed is proposed in this paper. ACSFed can dynamically calculate the possibility of clients being selected to train the model for each communication round based on their local statistical heterogeneity and previous training performance instead of randomly selected clients, and clients with heavier statistical heterogeneity or bad training performance would be more likely selected to participate in the later training. This client selection strategy can enable the federated model to learn the global statistical knowledge faster and thereby promote the convergence of the federated model. Multiple experiments on public benchmark datasets demonstrate these improvements in the efficiency of the models in heterogeneous settings.
引用
收藏
页数:14
相关论文
共 31 条
[1]  
[Anonymous], 2013, P 1 INT WORKSH MOB C
[2]   FedOpt: Towards Communication Efficiency and Privacy Preservation in Federated Learning [J].
Asad, Muhammad ;
Moustafa, Ahmed ;
Ito, Takayuki .
APPLIED SCIENCES-BASEL, 2020, 10 (08)
[3]  
Cai Z., IEEE T MOBILE COMPUT
[4]   Trading Private Range Counting over Big IoT Data [J].
Cai, Zhipeng ;
He, Zaobo .
2019 39TH IEEE INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS 2019), 2019, :144-153
[5]   A Private and Efficient Mechanism for Data Uploading in Smart Cyber-Physical Systems [J].
Cai, Zhipeng ;
Zheng, Xu .
IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2020, 7 (02) :766-775
[6]  
Chen W., 2020, arXiv
[7]  
Cho Y.J., 2020, ARXIV PREPRINT ARXIV, DOI [10.48550/arXiv.2010.01243, DOI 10.48550/ARXIV.2010.01243]
[8]   LoAdaBoost: Loss-based AdaBoost federated machine learning with reduced computational complexity on IID and non-IID intensive care data [J].
Huang, Li ;
Yin, Yifeng ;
Fu, Zeng ;
Zhang, Shifa ;
Deng, Hao ;
Liu, Dianbo .
PLOS ONE, 2020, 15 (04)
[9]  
Jeong E., 2018, ARXIV PREPRINT ARXIV, DOI [10.48550/arXiv.1811.11479, DOI 10.48550/ARXIV.1811.11479]
[10]  
Koneny J., 2016, ARXIV PREPRINT ARXIV, DOI [10.48550/arXiv.1610.05492, DOI 10.48550/ARXIV.1610.05492]