Parameter Communication Consistency Model for Large-Scale Security Monitoring Based on Mobile Computing

被引:2
作者
Yang, Rui [1 ,2 ]
Zhang, Jilin [1 ,2 ,3 ]
Wan, Jian [1 ,2 ,4 ]
Zhou, Li [1 ,2 ]
Shen, Jing [1 ,2 ]
Zhang, Yunchen [1 ,2 ]
Wei, Zhenguo [5 ]
Zhang, Juncong [5 ]
Wang, Jue [6 ]
机构
[1] Hangzhou Dianzi Univ, Sch Comp Sci & Technol, Hangzhou 310018, Peoples R China
[2] Hangzhou Dianzi Univ, Minist Educ, Key Lab Complex Syst Modeling & Simulat, Hangzhou 310018, Peoples R China
[3] Chinese Acad Sci, Inst Comp Technol, State Key Lab Comp Architecture, Beijing 100190, Peoples R China
[4] Zhejiang Univ Sci & Technol, Sch Informat & Elect Engn, Hangzhou 310023, Peoples R China
[5] Chinese Acad Sci, Comp Network Informat Ctr, Beijing 100190, Peoples R China
[6] Zhejiang Dawning Informat Technol Co Ltd, Hangzhou 310051, Peoples R China
基金
中国国家自然科学基金;
关键词
Mobile computing; security monitoring; distributed machine learning; limited synchronous parallel model; parameter server;
D O I
10.1109/ACCESS.2019.2956632
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the application of mobile computing in the security field, security monitoring big data has also begun to emerge, providing favorable support for smart city construction and city-scale and investment expansion. Mobile computing takes full advantage of the computing power and communication capabilities of various sensing devices and uses these devices to form a computing cluster. When using such clusters for training of distributed machine learning models, the load imbalance and network transmission delay result in low efficiency of model training. Therefore, this paper proposes a distributed machine learning parameter communication consistency model based on the parameter server idea, which is called the limited synchronous parallel model. The model is based on the fault-tolerant characteristics of the machine learning algorithm, and it dynamically limits the size of the synchronization barrier of the parameter server, reduces the synchronization communication overhead, and ensures the accuracy of the model training; thus, the model realizes finite asynchronous calculation between the worker nodes and gives full play to the overall performance of the cluster. The implementation of cluster dynamic load balancing experiments shows that the model can fully utilize the cluster performance during the training of distributed machine learning models to ensure the accuracy of the model and improve the training speed.
引用
收藏
页码:171884 / 171897
页数:14
相关论文
共 45 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]  
Ahmed A, 2012, 2012 4TH CONFERENCE ON DATA MINING AND OPTIMIZATION (DMO), P121, DOI 10.1109/DMO.2012.6329809
[3]  
[Anonymous], 2019, Journal of Ambient Intelligence and Humanized Computing, DOI DOI 10.1007/S12652-019-01223-3
[4]  
[Anonymous], 2013, Proceedings of the 30th International Conference on Machine Learning, Cycle
[5]  
[Anonymous], 2018, J MACHINE LEARNING R
[6]  
[Anonymous], SENSORS
[7]  
[Anonymous], 2015, P NEUR INF PROC SYST
[8]   Straggler Mitigation for Distributed Behavioral Simulation [J].
Bin Khunayn, Eman ;
Karunasekera, Shanika ;
Xie, Hairuo ;
Ramamohanarao, Kotagiri .
2017 IEEE 37TH INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS 2017), 2017, :2638-2641
[9]  
Chilimbi Trishul M, 2014, P USENIX OSDI, V14, P571
[10]  
Collobert R., 2011, BIGLEARN NIPS WORKSH, P1