A parallel and distributed stochastic gradient descent implementation using commodity clusters

被引:0
|
作者
Robert K. L. Kennedy
Taghi M. Khoshgoftaar
Flavio Villanustre
Timothy Humphrey
机构
[1] Florida Atlantic University,
[2] LexisNexis Business Information Solutions,undefined
来源
关键词
Parallel stochastic gradient descent; Parallel and distributed processing; Deep learning; Big data; Neural network; Cluster computer; HPCC systems;
D O I
暂无
中图分类号
学科分类号
摘要
Deep Learning is an increasingly important subdomain of artificial intelligence, which benefits from training on Big Data. The size and complexity of the model combined with the size of the training dataset makes the training process very computationally and temporally expensive. Accelerating the training process of Deep Learning using cluster computers faces many challenges ranging from distributed optimizers to the large communication overhead specific to systems with off the shelf networking components. In this paper, we present a novel distributed and parallel implementation of stochastic gradient descent (SGD) on a distributed cluster of commodity computers. We use high-performance computing cluster (HPCC) systems as the underlying cluster environment for the implementation. We overview how the HPCC systems platform provides the environment for distributed and parallel Deep Learning, how it provides a facility to work with third party open source libraries such as TensorFlow, and detail our use of third-party libraries and HPCC functionality for implementation. We provide experimental results that validate our work and show that our implementation can scale with respect to both dataset size and the number of compute nodes in the cluster.
引用
收藏
相关论文
共 50 条
  • [31] Two-beam coherent addition using the stochastic parallel gradient descent algorithm
    Yang, Yuchuan
    Tan, Bitao
    Lizi, Chen
    Zhang, Junwei
    Wang, Xiao
    OPTICAL ENGINEERING, 2013, 52 (01)
  • [32] Investigations of incoherent beam combining using stochastic parallel gradient descent with retroreflector target
    Henriksson, Markus
    Brannlund, Carl
    TECHNOLOGIES FOR OPTICAL COUNTERMEASURES XVI, 2019, 11161
  • [33] Performance of stochastic parallel gradient descent algorithm in coherent combination
    Li X.
    He Y.
    1600, Chinese Optical Society (36):
  • [34] Stochastic parallel gradient descent algorithm for adaptive optics system
    Ma H.
    Zhang P.
    Zhang J.
    Fan C.
    Wang Y.
    Qiangjiguang Yu Lizishu/High Power Laser and Particle Beams, 2010, 22 (06): : 1206 - 1210
  • [35] A Sharp Estimate on the Transient Time of Distributed Stochastic Gradient Descent
    Pu, Shi
    Olshevsky, Alex
    Paschalidis, Ioannis Ch
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (11) : 5900 - 5915
  • [36] A Distributed Optimal Control Problem with Averaged Stochastic Gradient Descent
    Sun, Qi
    Du, Qiang
    COMMUNICATIONS IN COMPUTATIONAL PHYSICS, 2020, 27 (03) : 753 - 774
  • [37] Scaling Stratified Stochastic Gradient Descent for Distributed Matrix Completion
    Abubaker N.
    Karsavuran M.O.
    Aykanat C.
    IEEE Transactions on Knowledge and Data Engineering, 2023, 35 (10) : 10603 - 10615
  • [38] Distributed Stochastic Gradient Descent with Event-Triggered Communication
    George, Jemin
    Gurram, Prudhvi
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 7169 - 7178
  • [39] ON DISTRIBUTED STOCHASTIC GRADIENT DESCENT FOR NONCONVEX FUNCTIONS IN THE PRESENCE OF BYZANTINES
    Bulusu, Saikiran
    Khanduri, Prashant
    Sharma, Pranay
    Varshney, Pramod K.
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3137 - 3141
  • [40] Distributed Differentially Private Stochastic Gradient Descent: An Empirical Study
    Hegedus, Istvan
    Jelasity, Mark
    2016 24TH EUROMICRO INTERNATIONAL CONFERENCE ON PARALLEL, DISTRIBUTED, AND NETWORK-BASED PROCESSING (PDP), 2016, : 566 - 573