Spatio-Temporal Pruning and Quantization for Low-latency Spiking Neural Networks

被引:18
|
作者
Chowdhury, Sayeed Shafayet [1 ]
Garg, Isha [1 ]
Roy, Kaushik [1 ]
机构
[1] Purdue Univ, Sch Elect & Comp Engn, W Lafayette, IN 47905 USA
基金
美国国家科学基金会;
关键词
SNN; temporal pruning; latency; spike rate; quantization; accuracy;
D O I
10.1109/IJCNN52387.2021.9534111
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Spiking Neural Networks (SNNs) are a promising alternative to traditional deep learning methods since they perform event-driven information processing. However, a major drawback of SNNs is high inference latency. The efficiency of SNNs could be enhanced using compression methods such as pruning and quantization. Notably, SNNs, unlike their non-spiking counterparts, consist of a temporal dimension, the compression of which can lead to latency reduction. In this paper, we propose spatial and temporal pruning of SNNs. First, structured spatial pruning is performed by determining the layer-wise significant dimensions using principal component analysis of the average accumulated membrane potential of the neurons. This step leads to 10-14X model compression. Additionally, it enables inference with lower latency and decreases the spike count per inference. To further reduce latency, temporal pruning is performed by gradually reducing the timesteps while training. The networks are trained using surrogate gradient descent based backpropagation and we validate the results on CIFAR10 and CIFAR100, using VGG architectures. The spatio-temporally pruned SNNs achieve 89.04% and 66.4% accuracy on CIFAR10 and CIFAR100, respectively, while performing inference with 3-30X reduced latency compared to state-of-the-art SNNs. Moreover, they require 8-14X lesser compute energy compared to their unpruned standard deep learning counterparts. The energy numbers are obtained by multiplying the number of operations with energy per operation. These SNNs also provide 1-4% higher robustness against Gaussian noise corrupted inputs. Furthermore, we perform weight quantization and find that performance remains reasonably stable up to 5-bit quantization.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Dynamic spatio-temporal pruning for efficient spiking neural networks
    Gou, Shuiping
    Fu, Jiahui
    Sha, Yu
    Cao, Zhen
    Guo, Zhang
    Eshraghian, Jason K.
    Li, Ruimin
    Jiao, Licheng
    FRONTIERS IN NEUROSCIENCE, 2025, 19
  • [2] Spatio-Temporal Pruning for Training Ultra-Low-Latency Spiking Neural Networks in Remote Sensing Scene Classification
    Li, Jiahao
    Xu, Ming
    Chen, He
    Liu, Wenchao
    Chen, Liang
    Xie, Yizhuang
    REMOTE SENSING, 2024, 16 (17)
  • [3] Ultra-low latency spiking neural networks with spatio-temporal compression and synaptic convolutional block
    Xu, Changqing
    Liu, Yi
    Yang, Yintang
    NEUROCOMPUTING, 2023, 550
  • [4] SiBrain: A Sparse Spatio-Temporal Parallel Neuromorphic Architecture for Accelerating Spiking Convolution Neural Networks With Low Latency
    Chen, Yuehai
    Ye, Wujian
    Liu, Yijun
    Zhou, Huihui
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2024, : 1 - 13
  • [5] Spatio-temporal Representations of Uncertainty in Spiking Neural Networks
    Savin, Cristina
    Deneve, Sophie
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014), 2014, 27
  • [6] Optimized Potential Initialization for Low-Latency Spiking Neural Networks
    Bu, Tong
    Ding, Jianhao
    Yu, Zhaofei
    Huang, Tiejun
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 11 - 20
  • [7] Can Deep Neural Networks be Converted to Ultra Low-Latency Spiking Neural Networks?
    Datta, Gourav
    Beerel, Peter A.
    PROCEEDINGS OF THE 2022 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2022), 2022, : 718 - 723
  • [8] Convolutional Spiking Neural Networks for Spatio-Temporal Feature Extraction
    Ali Samadzadeh
    Fatemeh Sadat Tabatabaei Far
    Ali Javadi
    Ahmad Nickabadi
    Morteza Haghir Chehreghani
    Neural Processing Letters, 2023, 55 : 6979 - 6995
  • [9] Convolutional Spiking Neural Networks for Spatio-Temporal Feature Extraction
    Samadzadeh, Ali
    Far, Fatemeh Sadat Tabatabaei
    Javadi, Ali
    Nickabadi, Ahmad
    Chehreghani, Morteza Haghir
    NEURAL PROCESSING LETTERS, 2023, 55 (06) : 6979 - 6995
  • [10] Constrain Bias Addition to Train Low-Latency Spiking Neural Networks
    Lin, Ranxi
    Dai, Benzhe
    Zhao, Yingkai
    Chen, Gang
    Lu, Huaxiang
    BRAIN SCIENCES, 2023, 13 (02)