Research on Selective Combination of Distributed Machine Learning Models

被引:0
作者
Tsuchiya T. [1 ,2 ]
Mochizuki R. [2 ]
Hirose H. [2 ]
Yamada T. [2 ]
Minh Q.T. [3 ]
机构
[1] Tokyo International University, Tokyo
[2] Suwa University of Science, Chino
[3] Ho Chi Minh City University of Technology (HCMUT), VNU-HCM, Ho Chi Minh City
关键词
Distributed machine learning; Fog computing; Learning model combination;
D O I
10.1007/s42979-022-01312-9
中图分类号
学科分类号
摘要
We have proposed a distributed platform for machine learning without data accumulation. This method constructs feature models from distributed data and combines them to obtain the same level of performance as conventional methods based on data accumulation. In this paper, we propose and compare three methods for selecting and combining these combinable feature models named fog model; sequential model selection, a conventional method to select fog models in the order of node IDs; adaptive selection, a method to select only models to improve task performance; and similar model selection, a method to select models similar to the current model. Each method has different priorities in selecting models. As an evaluation, the proposed methods are compared with the method of previous research by simulation. The proposed method which gives priority to performance improvement with small number of combinations for the processing tasks comparison to previous one. The proposed method enables more efficient combinations of fog models with fewer models than previous research, and users are able to adaptively select fog models to improve overall performance or to prioritize the performance of specific features by selecting the target models. © 2022, The Author(s), under exclusive licence to Springer Nature Singapore Pte Ltd.
引用
收藏
相关论文
共 14 条
[1]  
Tsuchiya T., Mochizuki R., Hirose H.Y., Koyanagi T.K., Quang T.M., Distributed data platform for machine learning using the fog computing model, SN Comput Sci, 1, (2020)
[2]  
Bonomi F., Milito R., Natarajan P., Zhu J., Fog computing: a platform for internet of things and analytics, big data and internet of things: a roadmap for smart environments, pp. 169-186, (2014)
[3]  
Laperdrix P., Bielova N., Baudry B., Avoine G., Browser fingerprinting: A survey, (2019)
[4]  
McMahan H.B., Moore E., Ramage D., Hampson S., Arcas B.A., Communication-efficient learning of deep networks from decentralized data, Proceeding of the 20Th Int’l Conference on Artificial Intelligence and Statistics, JMLR: W&CP, 54, pp. 169-186, (2014)
[5]  
Devlin J., Chang M.-W., Lee K., Toutanova K., BERT: Pre-training of deep bidirectional transformers for language understanding, (2018)
[6]  
Bonomi F., Milito R., Natarajan P., Zhu J., Bessis N., Dobre C., Fog computing: a platform for internet of things and analytics, big data and internet of things: a roadmap for smart environments, pp. 169-186, (2014)
[7]  
Zhou X., Huang S., Zheng Z., RPD: A distance function between word embeddings, Proceedings of 58Th the Association for Computational Linguistics, pp. 42-50, (2020)
[8]  
Agirre E., Alfonseca E., Hall K., Kravalova J., Pasca M., Soroa A., A study on similarity and relatedness using distributional and wordnet-based approaches, Proceedings of Human Language Technologies: The North American Chapter of The Association for Computational Linguistics, Boulder, pp. 19-27, (2009)
[9]  
(2022)
[10]  
Tomas M., Kai C., Greg C., Jeffrey D., Efficient estimation of word representations in vector space, (2013)