Input-Aware Dynamic Timestep Spiking Neural Networks for Efficient In-Memory Computing

被引:3
作者
Li, Yuhang [1 ]
Moitra, Ahhishek [1 ]
Geller, Tamar [1 ]
Panda, Priyadarshini [1 ]
机构
[1] Yale Univ, Dept Elect Engn, New Haven, CT 06511 USA
来源
2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC | 2023年
关键词
Spiking neural networks; in-memory computing; dynamic inference;
D O I
10.1109/DAC56929.2023.10247869
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Spiking Neural Networks (SNNs) have recently attracted widespread research interest as an efficient alternative to traditional Artificial Neural Networks (ANNs) because of their capability to process sparse and binary spike information and avoid expensive multiplication operations. Although the efficiency of SNNs can be realized on the In-Memory Computing (IMC) architecture, we show that the energy cost and latency of SNNs scale linearly with the number of timesteps used on IMC hardware. Therefore, in order to maximize the efficiency of SNNs, we propose input-aware Dynamic Timestep SNN (DT-SNN), a novel algorithmic solution to dynamically determine the number of timesteps during inference on an input-dependent basis. By calculating the entropy of the accumulated output after each timestep, we can compare it to a predefined threshold and decide if the information processed at the current timestep is sufficient for a confident prediction. We deploy DT-SNN on an IMC architecture and show that it incurs negligible computational overhead. We demonstrate that our method only uses 1.46 average timesteps to achieve the accuracy of a 4-timestep static SNN while reducing the energy-delay-product by 80%.
引用
收藏
页数:6
相关论文
共 23 条
  • [1] Bolukbasi T, 2017, PR MACH LEARN RES, V70
  • [2] NeuroSim: A Circuit-Level Macro Model for Benchmarking Neuro-Inspired Architectures in Online Learning
    Chen, Pai-Yu
    Peng, Xiaochen
    Yu, Shimeng
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2018, 37 (12) : 3067 - 3080
  • [3] Spatio-Temporal Pruning and Quantization for Low-latency Spiking Neural Networks
    Chowdhury, Sayeed Shafayet
    Garg, Isha
    Roy, Kaushik
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [4] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
  • [5] Guo CA, 2017, PR MACH LEARN RES, V70
  • [6] Han S, 2016, Arxiv, DOI [arXiv:1510.00149, DOI 10.48550/ARXIV.1510.00149]
  • [7] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [8] Revisiting Batch Normalization for Training Low-Latency Deep Spiking Neural Networks From Scratch
    Kim, Youngeun
    Panda, Priyadarshini
    [J]. FRONTIERS IN NEUROSCIENCE, 2021, 15
  • [9] Krizhevsky A, 2009, LEARNING MULTIPLE LA, P5
  • [10] Deep learning
    LeCun, Yann
    Bengio, Yoshua
    Hinton, Geoffrey
    [J]. NATURE, 2015, 521 (7553) : 436 - 444