Quantisation and pooling method for low-inference-latency spiking neural networks

被引:10
|
作者
Lin, Zhitao [1 ]
Shen, Juncheng [1 ]
Ma, De [2 ]
Meng, Jianyi [3 ]
机构
[1] Zhejiang Univ, Inst VLSI Design, Hangzhou, Zhejiang, Peoples R China
[2] Zhejiang Univ, Coll Comp Sci, Hangzhou, Zhejiang, Peoples R China
[3] Fudan Univ, State Key Lab ASIC & Syst, Shanghai, Peoples R China
关键词
neural nets; object recognition; real-time recognition tasks; CIFAR10; MNIST; spiking neurons; convolutional layers; pooling function; retraining; layer-wise quantisation method; DNN; deep neural network; SNN; low-inference-latency spiking neural networks; pooling method;
D O I
10.1049/el.2017.2219
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Spiking neural network (SNN) that converted from conventional deep neural network (DNN) has shown great potential as a solution for fast and efficient recognition. A layer-wise quantisation method based on retraining is proposed to quantise the activation of DNN, which reduces the number of time steps required by converted SNN to achieve minimal accuracy loss. Pooling function is incorporated into convolutional layers to reduce at most 20% of spiking neurons. The converted SNNs achieved 99.15% accuracy on MNIST and 82.9% on CIFAR10 by only seven time steps, and only 10-40% of spikes need to be processed compared with networks using traditional algorithms. The experimental results show that the proposed methods are able to build hardware-friendly SNNs with ultra-low-inference latency.
引用
收藏
页码:1347 / 1348
页数:2
相关论文
共 50 条
  • [41] Probabilistic Inference Using Stochastic Spiking Neural Networks on A Neurosynaptic Processor
    Ahmed, Khadeer
    Shrestha, Amar
    Qiu, Qinru
    Wu, Qing
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 4286 - 4293
  • [42] Simulation of Bayesian Learning and Inference on Distributed Stochastic Spiking Neural Networks
    Ahmed, Khadeer
    Shrestha, Amar
    Qiu, Qinru
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 1044 - 1051
  • [43] Ultra-low latency spiking neural networks with spatio-temporal compression and synaptic convolutional block
    Xu, Changqing
    Liu, Yi
    Yang, Yintang
    NEUROCOMPUTING, 2023, 550
  • [44] Direct Training via Backpropagation for Ultra-Low-Latency Spiking Neural Networks with Multi-Threshold
    Xu, Changqing
    Liu, Yi
    Chen, Dongdong
    Yang, Yintang
    SYMMETRY-BASEL, 2022, 14 (09):
  • [45] A universal ANN-to-SNN framework for achieving high accuracy and low latency deep Spiking Neural Networks
    Wang, Yuchen
    Liu, Hanwen
    Zhang, Malu
    Luo, Xiaoling
    Qu, Hong
    NEURAL NETWORKS, 2024, 174
  • [46] Toward High-Accuracy and Low-Latency Spiking Neural Networks With Two-Stage Optimization
    Wang, Ziming
    Zhang, Yuhao
    Lian, Shuang
    Cui, Xiaoxin
    Yan, Rui
    Tang, Huajin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, : 1 - 15
  • [47] Low-Latency Spiking Neural Networks Using Pre-Charged Membrane Potential and Delayed Evaluation
    Hwang, Sungmin
    Chang, Jeesoo
    Oh, Min-Hye
    Min, Kyung Kyu
    Jang, Taejin
    Park, Kyungchul
    Yu, Junsu
    Lee, Jong-Ho
    Park, Byung-Gook
    FRONTIERS IN NEUROSCIENCE, 2021, 15
  • [48] A Low Power and Low Latency FPGA-Based Spiking Neural Network Accelerator
    Liu, Hanwen
    Chen, Yi
    Zeng, Zihang
    Zhang, Malu
    Qu, Hong
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [49] GAZELLE: A Low Latency Framework for Secure Neural Network Inference
    Juvekar, Chiraag
    Vaikuntanathan, Vinod
    Chandrakasan, Anantha
    PROCEEDINGS OF THE 27TH USENIX SECURITY SYMPOSIUM, 2018, : 1651 - 1668
  • [50] Bridging the Gap between Spiking Neural Networks & LSTMs for Latency & Energy Efficiency
    Datta, Gourav
    Deng, Haoqin
    Aviles, Robert
    Liu, Zeyu
    Beerel, Peter A.
    2023 IEEE/ACM INTERNATIONAL SYMPOSIUM ON LOW POWER ELECTRONICS AND DESIGN, ISLPED, 2023,