Can Deep Neural Networks be Converted to Ultra Low-Latency Spiking Neural Networks?

被引:0
|
作者
Datta, Gourav [1 ]
Beerel, Peter A. [1 ]
机构
[1] Univ Southern Calif, Ming Hsieh Dept Elect & Comp Engn, Los Angeles, CA 90089 USA
关键词
SNN; DNN; neuromorphic; FLOPs; surrogate gradient learning;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Spiking neural networks (SNNs), that operate via binary spikes distributed over time, have emerged as a promising energy efficient ML paradigm for resource-constrained devices. However, the current state-of-the-art (SOTA) SNNs require multiple time steps for acceptable inference accuracy, increasing spiking activity and, consequently, energy consumption. SOTA training strategies for SNNs involve conversion from a non-spiking deep neural network (DNN). In this paper, we determine that SOTA conversion strategies cannot yield ultra low latency because they incorrectly assume that the DNN and SNN pre-activation values are uniformly distributed. We propose a new training algorithm that accurately captures these distributions, minimizing the error between the DNN and converted SNN. The resulting SNNs have ultra low latency and high activation sparsity, yielding significant improvements in compute efficiency. In particular, we evaluate our framework on image recognition tasks from CIFAR-10 and CIFAR-100 datasets on several VGG and ResNet architectures. We obtain top-1 accuracy of 64.19% with only 2 time steps on the CIFAR-100 dataset with similar to 159.2x lower compute energy compared to an iso-architecture standard DNN. Compared to other SOTA SNN models, our models perform inference 2.5-8x faster (i.e., with fewer time steps).
引用
收藏
页码:718 / 723
页数:6
相关论文
共 50 条
  • [41] Neuromorphic Architectures for Spiking Deep Neural Networks
    Indiveri, Giacomo
    Corradi, Federico
    Qiao, Ning
    2015 IEEE INTERNATIONAL ELECTRON DEVICES MEETING (IEDM), 2015,
  • [42] Deep Residual Learning in Spiking Neural Networks
    Fang, Wei
    Yu, Zhaofei
    Chen, Yanqi
    Huang, Tiejun
    Masquelier, Timothee
    Tian, Yonghong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [43] Constructing Deep Spiking Neural Networks from Artificial Neural Networks with Knowledge Distillation
    Xu, Qi
    Li, Yaxin
    Shen, Jiangrong
    Liu, Jian K.
    Tang, Huajin
    Pan, Gang
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 7886 - 7895
  • [44] A universal ANN-to-SNN framework for achieving high accuracy and low latency deep Spiking Neural Networks
    Wang, Yuchen
    Liu, Hanwen
    Zhang, Malu
    Luo, Xiaoling
    Qu, Hong
    NEURAL NETWORKS, 2024, 174
  • [45] Deep Neural Network Supervised Bandwidth Allocation Decisions for Low-Latency Heterogeneous E-Health Networks
    Ruan, Lihua
    Dias, Maluge Pubuduni Imali
    Wong, Elaine
    JOURNAL OF LIGHTWAVE TECHNOLOGY, 2019, 37 (16) : 4147 - 4154
  • [46] Third Generation Neural Networks: Spiking Neural Networks
    Ghosh-Dastidar, Samanwoy
    Adeli, Hojjat
    ADVANCES IN COMPUTATIONAL INTELLIGENCE, 2009, 61 : 167 - +
  • [47] SPIKING NEURAL NETWORKS
    Ghosh-Dastidar, Samanwoy
    Adeli, Hojjat
    INTERNATIONAL JOURNAL OF NEURAL SYSTEMS, 2009, 19 (04) : 295 - 308
  • [48] Dynamic threshold integrate and fire neuron model for low latency spiking neural networks
    Wu, Xiyan
    Zhao, Yufei
    Song, Yong
    Jiang, Yurong
    Bai, Yashuo
    Li, Xinyi
    Zhou, Ya
    Yang, Xin
    Hao, Qun
    NEUROCOMPUTING, 2023, 544
  • [49] Scalable Energy-Efficient, Low-Latency Implementations of Trained Spiking Deep Belief Networks on SpiNNaker
    Stromatias, Evangelos
    Neil, Daniel
    Galluppi, Francesco
    Pfeiffer, Michael
    Liu, Shih-Chii
    Furber, Steve
    2015 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2015,
  • [50] Applications of Deep Neural Networks for Ultra Low Power IoT
    Kodali, Sreela
    Hansen, Patrick
    Mulholland, Niamh
    Whatmough, Paul
    Brooks, David
    Wei, Gu-Yeon
    2017 IEEE 35TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD), 2017, : 589 - 592