Stochastic Markov gradient descent and training low-bit neural networks

被引:0
|
作者
Ashbrock, Jonathan [1 ]
Powell, Alexander M. [2 ]
机构
[1] MITRE Corp, Mclean, VA 22102 USA
[2] Vanderbilt Univ, Dept Math, Nashville, TN 37240 USA
来源
SAMPLING THEORY SIGNAL PROCESSING AND DATA ANALYSIS | 2021年 / 19卷 / 02期
关键词
Neural networks; Quantization; Stochastic gradient descent; Stochastic Markov gradient descent; Low-memory training;
D O I
10.1007/s43670-021-00015-1
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
The massive size of modern neural networks has motivated substantial recent interest in neural network quantization, especially low-bit quantization. We introduce Stochastic Markov Gradient Descent (SMGD), a discrete optimization method applicable to training quantized neural networks. The SMGD algorithm is designed for settings where memory is highly constrained during training. We provide theoretical guarantees of algorithm performance as well as encouraging numerical results.
引用
收藏
页数:23
相关论文
共 50 条
  • [41] Distributed stochastic gradient descent for link prediction in signed social networks
    Zhang, Han
    Wu, Gang
    Ling, Qing
    EURASIP JOURNAL ON ADVANCES IN SIGNAL PROCESSING, 2019, 2019 (1)
  • [42] An Efficient Stochastic Gradient Descent Algorithm to Maximize the Coverage of Cellular Networks
    Liu, Yaxi
    Wei Huangfu
    Zhang, Haijun
    Long, Keping
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2019, 18 (07) : 3424 - 3436
  • [43] Distributed stochastic gradient descent for link prediction in signed social networks
    Han Zhang
    Gang Wu
    Qing Ling
    EURASIP Journal on Advances in Signal Processing, 2019
  • [44] QUANTIZATION AND TRAINING OF LOW BIT-WIDTH CONVOLUTIONAL NEURAL NETWORKS FOR OBJECT DETECTION
    Yin, Penghang
    Zhang, Shuai
    Qi, Yingyong
    Xin, Jack
    JOURNAL OF COMPUTATIONAL MATHEMATICS, 2019, 37 (03) : 349 - 360
  • [45] On Stochastic Roundoff Errors in Gradient Descent with Low-Precision Computation
    Lu Xia
    Stefano Massei
    Michiel E. Hochstenbach
    Barry Koren
    Journal of Optimization Theory and Applications, 2024, 200 : 634 - 668
  • [46] On Stochastic Roundoff Errors in Gradient Descent with Low-Precision Computation
    Xia, Lu
    Massei, Stefano
    Hochstenbach, Michiel E.
    Koren, Barry
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2024, 200 (02) : 634 - 668
  • [47] Understanding and Optimizing Asynchronous Low-Precision Stochastic Gradient Descent
    De Sa, Christopher
    Feldman, Matthew
    Re, Christopher
    Olukotun, Kunle
    44TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA 2017), 2017, : 561 - 574
  • [48] Training Two-Layer ReLU Networks with Gradient Descent is Inconsistent
    Holzmueller, David
    Steinwart, Ingo
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [49] Forecasting the productivity of a solar distiller enhanced with an inclined absorber plate using stochastic gradient descent in artificial neural networks
    Mohammed, Suha A.
    Al-Haddad, Luttfi A.
    Alawee, Wissam H.
    Dhahad, Hayder A.
    Jaber, Alaa Abdulhady
    Al-Haddad, Sinan A.
    MULTISCALE AND MULTIDISCIPLINARY MODELING EXPERIMENTS AND DESIGN, 2024, 7 (03) : 1819 - 1829
  • [50] A hybrid model of heuristic algorithm and gradient descent to optimize neural networks
    Mirkhan, Amer
    Celebi, Numan
    BULLETIN OF THE POLISH ACADEMY OF SCIENCES-TECHNICAL SCIENCES, 2023, 71 (06)