DLBench: a comprehensive experimental evaluation of deep learning frameworks

被引:34
作者
Elshawi, Radwa [1 ]
Wahab, Abdul [1 ]
Barnawi, Ahmed [3 ]
Sakr, Sherif [2 ]
机构
[1] Univ Tartu, Tartu, Estonia
[2] Univ Tartu, Inst Comp Sci, Data Syst Grp, Tartu, Estonia
[3] King Abdulaziz Univ, Jeddah, Saudi Arabia
来源
CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS | 2021年 / 24卷 / 03期
关键词
Deep learning; Experimental evaluation; CNN; LSTM; ARCHITECTURES;
D O I
10.1007/s10586-021-03240-4
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Learning (DL) has achieved remarkable progress over the last decade on various tasks such as image recognition, speech recognition, and natural language processing. In general, three main crucial aspects fueled this progress: the increasing availability of large amount of digitized data, the increasing availability of affordable parallel and powerful computing resources (e.g., GPU) and the growing number of open source deep learning frameworks that facilitate and ease the development process of deep learning architectures. In practice, the increasing popularity of deep learning frameworks calls for benchmarking studies that can effectively evaluate and understand the performance characteristics of these systems. In this paper, we conduct an extensive experimental evaluation and analysis of six popular deep learning frameworks, namely, TensorFlow, MXNet, PyTorch, Theano, Chainer, and Keras, using three types of DL architectures Convolutional Neural Networks (CNN), Faster Region-based Convolutional Neural Networks (Faster R-CNN), and Long Short Term Memory (LSTM). Our experimental evaluation considers different aspects for its comparison including accuracy, training time, convergence and resource consumption patterns. Our experiments have been conducted on both CPU and GPU environments using different datasets. We report and analyze the performance characteristics of the studied frameworks. In addition, we report a set of insights and important lessons that we have learned from conducting our experiments.
引用
收藏
页码:2017 / 2038
页数:22
相关论文
共 53 条
[11]  
Bastien F., 2012, ARXIV PREPRINT ARXIV
[12]   Learning Deep Architectures for AI [J].
Bengio, Yoshua .
FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2009, 2 (01) :1-127
[13]   Equivalence among Stochastic Logic Circuits and its Application [J].
Chen, Te-Hsuan ;
Hayes, John P. .
2015 52ND ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2015,
[14]  
Chollet F, 2015, Keras: The Python Deep Learning library
[15]  
Coleman C., 2017, Training, V100, P102
[16]  
Collobert R., 2011, BIGLEARN NIPS WORKSH
[17]  
Collobert R, 2011, J MACH LEARN RES, V12, P2493
[18]   OpenMP: An industry standard API for shared-memory programming [J].
Dagum, L ;
Menon, R .
IEEE COMPUTATIONAL SCIENCE & ENGINEERING, 1998, 5 (01) :46-55
[19]   The PASCAL Visual Object Classes Challenge: A Retrospective [J].
Everingham, Mark ;
Eslami, S. M. Ali ;
Van Gool, Luc ;
Williams, Christopher K. I. ;
Winn, John ;
Zisserman, Andrew .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2015, 111 (01) :98-136
[20]   Interference-aware parallelization for deep learning workload in GPU cluster [J].
Geng, Xin ;
Zhang, Haitao ;
Zhao, Zhengyang ;
Ma, Huadong .
CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2020, 23 (04) :2689-2702