Quantisation and pooling method for low-inference-latency spiking neural networks

被引:10
|
作者
Lin, Zhitao [1 ]
Shen, Juncheng [1 ]
Ma, De [2 ]
Meng, Jianyi [3 ]
机构
[1] Zhejiang Univ, Inst VLSI Design, Hangzhou, Zhejiang, Peoples R China
[2] Zhejiang Univ, Coll Comp Sci, Hangzhou, Zhejiang, Peoples R China
[3] Fudan Univ, State Key Lab ASIC & Syst, Shanghai, Peoples R China
关键词
neural nets; object recognition; real-time recognition tasks; CIFAR10; MNIST; spiking neurons; convolutional layers; pooling function; retraining; layer-wise quantisation method; DNN; deep neural network; SNN; low-inference-latency spiking neural networks; pooling method;
D O I
10.1049/el.2017.2219
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Spiking neural network (SNN) that converted from conventional deep neural network (DNN) has shown great potential as a solution for fast and efficient recognition. A layer-wise quantisation method based on retraining is proposed to quantise the activation of DNN, which reduces the number of time steps required by converted SNN to achieve minimal accuracy loss. Pooling function is incorporated into convolutional layers to reduce at most 20% of spiking neurons. The converted SNNs achieved 99.15% accuracy on MNIST and 82.9% on CIFAR10 by only seven time steps, and only 10-40% of spikes need to be processed compared with networks using traditional algorithms. The experimental results show that the proposed methods are able to build hardware-friendly SNNs with ultra-low-inference latency.
引用
收藏
页码:1347 / 1348
页数:2
相关论文
共 50 条
  • [21] Training Low-Latency Spiking Neural Network with Orthogonal Spiking Neurons
    Yao, Yunpeng
    Wu, Man
    Zhang, Renyuan
    2023 21ST IEEE INTERREGIONAL NEWCAS CONFERENCE, NEWCAS, 2023,
  • [22] Converting Artificial Neural Networks to Ultralow-Latency Spiking Neural Networks for Action Recognition
    You, Hong
    Zhong, Xian
    Liu, Wenxuan
    Wei, Qi
    Huang, Wenxin
    Yu, Zhaofei
    Huang, Tiejun
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2024, 16 (04) : 1533 - 1545
  • [23] A Low Latency Spiking Neural Network with Improved Temporal Dynamics
    Yao, Yunpeng
    Kan, Yirong
    Zhu, Guangxian
    Zhang, Renyuan
    2023 IEEE 36TH INTERNATIONAL SYSTEM-ON-CHIP CONFERENCE, SOCC, 2023, : 226 - 231
  • [24] Partitioning Deep Neural Networks for Optimally Pipelined Inference on Heterogeneous IoT Devices with Low Latency Networks
    Sect, Woobean
    Kim, Saehwa
    Hong, Seongsoo
    2024 IEEE 44TH INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS, ICDCS 2024, 2024, : 1470 - 1471
  • [25] Accelerated Physical Emulation of Bayesian Inference in Spiking Neural Networks
    Kungl, Akos F.
    Schmitt, Sebastian
    Klaehn, Johann
    Mueller, Paul
    Baumbach, Andreas
    Dold, Dominik
    Kugele, Alexander
    Mueller, Eric
    Koke, Christoph
    Kleider, Mitja
    Mauch, Christian
    Breitwieser, Oliver
    Leng, Luziwei
    Guertler, Nico
    Guettler, Maurice
    Husmann, Dan
    Husmann, Kai
    Hartel, Andreas
    Karasenko, Vitali
    Gruebl, Andreas
    Schemmel, Johannes
    Meier, Karlheinz
    Petrovici, Mihai A.
    FRONTIERS IN NEUROSCIENCE, 2019, 13
  • [26] Training High-Performance Low-Latency Spiking Neural Networks by Differentiation on Spike Representation
    Meng, Qingyan
    Xiao, Mingqing
    Yan, Shen
    Wang, Yisen
    Lin, Zhouchen
    Luo, Zhi-Quan
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 12434 - 12443
  • [27] Towards Ultra Low Latency Spiking Neural Networks for Vision and Sequential Tasks Using Temporal Pruning
    Chowdhury, Sayeed Shafayet
    Rathi, Nitin
    Roy, Kaushik
    COMPUTER VISION, ECCV 2022, PT XI, 2022, 13671 : 709 - 726
  • [28] Revisiting Batch Normalization for Training Low-Latency Deep Spiking Neural Networks From Scratch
    Kim, Youngeun
    Panda, Priyadarshini
    FRONTIERS IN NEUROSCIENCE, 2021, 15
  • [29] CS-QCFS: Bridging the performance gap in ultra-low latency spiking neural networks
    Yang, Hongchao
    Yang, Suorong
    Zhang, Lingming
    Dou, Hui
    Shen, Furao
    Zhao, Jian
    NEURAL NETWORKS, 2025, 184
  • [30] Low Latency and Sparse Computing Spiking Neural Networks With Self-Driven Adaptive Threshold Plasticity
    Zhang, Anguo
    Shi, Jieming
    Wu, Junyi
    Zhou, Yongcheng
    Yu, Wei
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (12) : 17177 - 17188