Stochastic Computing Architectures for Lightweight LSTM Neural Networks

被引:0
作者
Sengupta, Roshwin [1 ]
Polian, Ilia [1 ]
Hayes, John P. [2 ]
机构
[1] Univ Stuttgart, Inst Comp Architecture & Comp Engn, Stuttgart, Germany
[2] Univ Michigan, Comp Engn Lab, Ann Arbor, MI 48109 USA
来源
2022 25TH INTERNATIONAL SYMPOSIUM ON DESIGN AND DIAGNOSTICS OF ELECTRONIC CIRCUITS AND SYSTEMS (DDECS) | 2022年
基金
美国国家科学基金会;
关键词
LSTM; recurrent neural nets; stochastic; computing;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
For emerging edge and near-sensor systems to perform hard classification tasks locally, they must avoid costly communication with the cloud. This requires the use of compact classifiers such as recurrent neural networks of the long short term memory (LSTM) type, as well as a low-area hardware technology such as stochastic computing (SC). We study the benefits and costs of applying SC to LSTM design. We consider a design space spanned by fully binary ( non-stochastic), fully stochastic, and several hybrid (mixed) LSTM architectures, and design and simulate examples of each. Using standard classification benchmarks, we show that area and power can be reduced up to 47% and 86% respectively with little or no impact on classification accuracy. We demonstrate that fully stochastic LSTMs can deliver acceptable accuracy despite accumulated errors. Our results also suggest that ReLU is preferable to tanh as an activation function in stochastic LSTMs
引用
收藏
页码:124 / 129
页数:6
相关论文
共 22 条
  • [1] The Promise and Challenge of Stochastic Computing
    Alaghi, Armin
    Qian, Weikang
    Hayes, John P.
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2018, 37 (08) : 1515 - +
  • [2] [Anonymous], 2019, ARXIV190202181
  • [3] LEARNING LONG-TERM DEPENDENCIES WITH GRADIENT DESCENT IS DIFFICULT
    BENGIO, Y
    SIMARD, P
    FRASCONI, P
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (02): : 157 - 166
  • [4] Stochastic neural computation I: Computational elements
    Brown, BD
    Card, HC
    [J]. IEEE TRANSACTIONS ON COMPUTERS, 2001, 50 (09) : 891 - 905
  • [5] Chang YN, 2013, INT CONF ACOUST SPEE, P2697, DOI 10.1109/ICASSP.2013.6638146
  • [6] Duch W., 1999, Neural Computing Surveys, V2
  • [7] Glorot X, 2011, P 14 INT C ART INT S, V15, P315, DOI DOI 10.1002/ECS2.1832
  • [8] Comparison of new activation functions in neural network for forecasting financial time series
    Gomes, Gecynalda S. da S.
    Ludermir, Teresa B.
    Lima, Leyla M. M. R.
    [J]. NEURAL COMPUTING & APPLICATIONS, 2011, 20 (03) : 417 - 439
  • [9] Fault latencies of concurrent checking FSMs
    Goot, R
    Levin, I
    Ostanin, S
    [J]. EUROMICRO SYMPOSIUM ON DIGITAL SYSTEM DESIGN, PROCEEDINGS: ARCHITECTURES, METHODS AND TOOLS, 2002, : 174 - 179
  • [10] Graves A, 2013, INT CONF ACOUST SPEE, P6645, DOI 10.1109/ICASSP.2013.6638947