Accelerating Deep Learning Systems via Critical Set Identification and Model Compression

被引:11
作者
Han, Rui [1 ]
Liu, Chi Harold [1 ]
Li, Shilin [1 ]
Wen, Shilin [1 ]
Liu, Xue [2 ]
机构
[1] Beijing Inst Technol, Sch Comp Sci & Technol, Beijing 100811, Peoples R China
[2] McGill Univ, Sch Comp Sci, Montreal, PQ H3A 0G4, Canada
基金
中国国家自然科学基金;
关键词
Training; Data models; Computational modeling; Acceleration; Synchronization; Monte Carlo methods; Deep learning; massive datasets; distributed systems; redundant input data;
D O I
10.1109/TC.2020.2970917
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Modern distributed engines are increasingly deployed to accelerate large-scaled deep learning (DL) training jobs. While the parallelism of distributed workers/nodes promises the scalability, the computation and communication overheads of the underlying iterative solving algorithms, e.g., stochastic gradient decent, unfortunately become the bottleneck for distributed DL training jobs. Existing approaches address such limitations by designing more efficient synchronization algorithms and model compressing techniques, but do not adequately address issues relating to processing massive datasets. In this article, we propose ClipDL, which accelerates the deep learning systems by simultaneously decreasing the number of model parameters as well as reducing the computations on critical data only. The core component of ClipDL is the estimation of critical set based on the observation that large proportions of input data have little influence on model parameter updating in many prevalent DL algorithms. We implemented ClipDL on Spark (a popular distributed engine for big data) and BigDL (based on de-factor distributed DL training architecture, parameter server), and integrated it with representative model compression techniques. The exhaustive experiments on real DL applications and datasets show ClipDL accelerates model training process by an average of 2.32 times while only incurring accuracy losses of 1.86 percent.
引用
收藏
页码:1059 / 1070
页数:12
相关论文
共 57 条
[41]  
Loshchilov I., 2016, PROC INT CONF LEARN
[42]  
Mu, 2014, Proceedings of the 11th USENIX Symposium on Operating Systems Design and Implementation (OSDI '14). OSDI '14, P583
[43]   XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks [J].
Rastegari, Mohammad ;
Ordonez, Vicente ;
Redmon, Joseph ;
Farhadi, Ali .
COMPUTER VISION - ECCV 2016, PT IV, 2016, 9908 :525-542
[44]   Neural-network feature selector [J].
Setiono, R ;
Liu, H .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1997, 8 (03) :654-662
[45]   Learning to Rank Short Text Pairs with Convolutional Deep Neural Networks [J].
Severyn, Aliaksei ;
Moschitti, Alessandro .
SIGIR 2015: PROCEEDINGS OF THE 38TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2015, :373-382
[46]  
Shokri R, 2015, ANN ALLERTON CONF, P909, DOI 10.1109/ALLERTON.2015.7447103
[47]  
Simonyan K, 2015, Arxiv, DOI arXiv:1409.1556
[48]  
Smith V., 2017, The Journal of Machine Learning Research, V18, P8590
[49]  
Song H., 2015, FIBER, P1, DOI DOI 10.48550/ARXIV.1510.00149
[50]  
Thierens D., 1993, Artificial Neural Nets and Genetic Algorithms. Proceedings of the International Conference, P658