DLBench: a comprehensive experimental evaluation of deep learning frameworks

被引:0
作者
Radwa Elshawi
Abdul Wahab
Ahmed Barnawi
Sherif Sakr
机构
[1] University of Tartu,
[2] King Abdulaziz University,undefined
来源
Cluster Computing | 2021年 / 24卷
关键词
Deep learning; Experimental evaluation; CNN; LSTM;
D O I
暂无
中图分类号
学科分类号
摘要
Deep Learning (DL) has achieved remarkable progress over the last decade on various tasks such as image recognition, speech recognition, and natural language processing. In general, three main crucial aspects fueled this progress: the increasing availability of large amount of digitized data, the increasing availability of affordable parallel and powerful computing resources (e.g., GPU) and the growing number of open source deep learning frameworks that facilitate and ease the development process of deep learning architectures. In practice, the increasing popularity of deep learning frameworks calls for benchmarking studies that can effectively evaluate and understand the performance characteristics of these systems. In this paper, we conduct an extensive experimental evaluation and analysis of six popular deep learning frameworks, namely, TensorFlow, MXNet, PyTorch, Theano, Chainer, and Keras, using three types of DL architectures Convolutional Neural Networks (CNN), Faster Region-based Convolutional Neural Networks (Faster R-CNN), and Long Short Term Memory (LSTM). Our experimental evaluation considers different aspects for its comparison including accuracy, training time, convergence and resource consumption patterns. Our experiments have been conducted on both CPU and GPU environments using different datasets. We report and analyze the performance characteristics of the studied frameworks. In addition, we report a set of insights and important lessons that we have learned from conducting our experiments.
引用
收藏
页码:2017 / 2038
页数:21
相关论文
共 41 条
  • [1] Abadi M(2016)Tensorflow: a system for large-scale machine learning OSDI 16 265-283
  • [2] Abugabah A(2020)Data mining techniques for analyzing healthcare conditions of urban space-person lung using meta-heuristic optimized neural networks Clust. Comput. 23 1781-1794
  • [3] AlZubi AA(2009)Learning deep architectures for AI Found. Trends Mach. Learn. 2 1-127
  • [4] Al-Obeidat FN(2011)Natural language processing (almost) from scratch J. Mach. Learn. Res. 12 2493-2537
  • [5] Alarifi A(1998)Openmp: an industry-standard API for shared-memory programming Comput. Sci. Eng. 1 46-55
  • [6] Alwadain A(2015)The pascal visual object classes challenge: a retrospective Int. J. Comput. Vis. 111 98-136
  • [7] Bengio Y(2020)Interference-aware parallelization for deep learning workload in GPU cluster Clust. Comput. 23 2689-2702
  • [8] Collobert R(2011)Comparison of values of pearson’s and spearman’s correlation coefficients on the same sets of data Quaestiones Geographicae 30 87-93
  • [9] Dagum L(2008)Amdahl’s law in the multicore era Computer 41 33-38
  • [10] Menon R(2012)Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups IEEE Signal Process. Mag. 29 82-97