A Non-deterministic Training Approach for Memory-Efficient Stochastic Neural Networks

被引:0
作者
Golbabaei, Babak [1 ]
Zhu, Guangxian [1 ]
Kan, Yirong [1 ]
Zhang, Renyuan [1 ]
Nakashima, Yasuhiko [1 ]
机构
[1] Nara Inst Sci & Technol, Div Informat Sci, Ikoma, Japan
来源
2023 IEEE 36TH INTERNATIONAL SYSTEM-ON-CHIP CONFERENCE, SOCC | 2023年
关键词
Non-determinism; Neural Network; Stochastic Encoding; Image Classification; Machine Learning; Voting Mechanism;
D O I
10.1109/SOCC58585.2023.10256838
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
We proposed a non-deterministic training approach for memory-efficient stochastic computing neural networks (SCNN). Conventional SCNN simply convert the trained network parameters into stochastic bit-streams at the inference phase. Although stochastic bit-streams can simplify binary multiplication and addition using the principle of probability, memory cost and computation delay are consumed due to extremely long bit-streams. Different from methods that rely on long bit-streams to convert full-precision NN to SCNN during inference, the proposed approach also introduces the concept of non-deterministic computation during training to alleviate the memory requirement growth caused by long bit-streams. To this end, we probabilize NN parameters in the feed-forward process of the training phase, and convert them into 1/4/8-bit stochastic number representations according to the probability, which greatly reduces the memory requirements in SC. In order to alleviate the training instability problem caused by low-bit encoding, we propose a multiple parallel training strategy (MPTS) during the training process to improve the stability of the results. The proposed MPTS achieves a stable training process through a voting mechanism. We evaluate the performance of the proposed training approach on fully connected NN and the MNIST dataset. Compared with the baseline training method with 97.77% accuracy using 32-bit floating point values, the proposed non-deterministic training approach achieves a reasonable accuracy of 90.34% while using 4-bit stochastic number representations to represent layer weights and biases.
引用
收藏
页码:232 / 237
页数:6
相关论文
共 27 条
[1]  
Alaghi A., 2012, ACM Transactions on Embedded Computing Systems, V12, P1
[2]  
Alaghi A, 2013, DES AUT CON
[3]  
Khan DA, 2019, Arxiv, DOI arXiv:1905.10906
[4]  
Blundell C, 2015, PR MACH LEARN RES, V37, P1613
[5]   A New Stochastic Computing Methodology for Efficient Neural Network Implementation [J].
Canals, Vincent ;
Morro, Antoni ;
Oliver, Antoni ;
Alomar, Miquel L. ;
Rossello, Josep L. .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2016, 27 (03) :551-564
[6]  
Chengguang Ma, 2012, 2012 International Conference on Computer Science and Service System (CSSS), P1587, DOI 10.1109/CSSS.2012.397
[7]  
Courbariaux M, 2016, Arxiv, DOI arXiv:1602.02830
[8]  
Dhillon G.S., 2018, arXiv, DOI DOI 10.48550/ARXIV.1803.01442
[9]  
Gaines B. R., 1969, Stochastic Computing Systems, V2, P37
[10]  
Gal Y, 2016, Arxiv, DOI [arXiv:1506.02142, 10.48550/arXiv.1506.02142]