GSP Distributed Deep Learning Used for the Monitoring System

被引:0
作者
Pan, Zhongming [1 ]
Luo, Yigui [1 ]
Sha, Wei [1 ]
Xie, Yin [1 ]
机构
[1] Tongji Univ, Sch Software Engn, Shanghai, Peoples R China
来源
2021 IEEE 6TH INTERNATIONAL CONFERENCE ON BIG DATA ANALYTICS (ICBDA 2021) | 2021年
关键词
distributed deep learning; data parallel; big data; communication mechanism;
D O I
10.1109/ICBDA51983.2021.9402960
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Monitoring system will collect massive data and store them in many distributed data centers. The traditional distributed deep learning needs to centralize data to the Parameter Server before training, then shuffle the data, and finally divide the data to each computing node uniformly. When training a neural network model with the data from monitoring system, it will result in huge communication overhead. We propose the Grouped-Based-on-Local-Data Synchronous Parallel (GSP) distributed deep learning. The use of local data can greatly reduce the amount of communication. At the same time, it can also bring another advantage, the similarity between the data centers can be calculated based on the background of monitor lens and the annotation of the data, then group computing nodes according to the similarity of the data centers they belong to and set up the grouped synchronous communication mechanism based on the Parameter Server architecture. The result of experiment shows that the iteration time of training by GSP is greatly reduced and the test precision is consistent.
引用
收藏
页码:224 / 229
页数:6
相关论文
共 16 条
  • [1] Benesty J, 2009, SPRINGER TOP SIGN PR, V2, P37, DOI 10.1007/978-3-642-00296-0_5
  • [2] Bochkovskiy A., ARXIV PREPRINT ARXIV
  • [3] Bradley J K, 2011, ARXIV PREPRINT ARXIV
  • [4] Chen J., 2016, ARXIV PREPRINT ARXIV
  • [5] Dean J., 2012, NIPS, V2012, P1223, DOI [10.5555/2999134.2999271, DOI 10.5555/2999134.2999271]
  • [6] Dekel O, 2012, J MACH LEARN RES, V13, P165
  • [7] Li M., 2014, 11 USENIX S OP SYST, P583
  • [8] Focal Loss for Dense Object Detection
    Lin, Tsung-Yi
    Goyal, Priya
    Girshick, Ross
    He, Kaiming
    Dollar, Piotr
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (02) : 318 - 327
  • [9] Convergence analysis of distributed stochastic gradient descent with shuffling
    Meng, Qi
    Chen, Wei
    Wang, Yue
    Ma, Zhi-Ming
    Liu, Tie-Yan
    [J]. NEUROCOMPUTING, 2019, 337 : 46 - 57
  • [10] Simonyan K, 2015, Arxiv, DOI arXiv:1409.1556