SSTDP: Supervised Spike Timing Dependent Plasticity for Efficient Spiking Neural Network Training

被引:30
作者
Liu, Fangxin [1 ,2 ]
Zhao, Wenbo [2 ,3 ]
Chen, Yongbiao [1 ]
Wang, Zongwu [1 ]
Yang, Tao [1 ]
Jiang, Li [1 ,2 ,4 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Elect Informat & Elect Engn, Shanghai, Peoples R China
[2] Shanghai Qi Zhi Inst, Shanghai, Peoples R China
[3] Columbia Univ, Sch Engn & Appl Sci, New York, NY USA
[4] Shanghai Jiao Tong Univ, Al Inst, MoE Key Lab Artificial Intelligence, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
spiking neural network; gradient descent backpropagation; neuromorphic computing; spike-time-dependent plasticity; deep learning; efficient training; BACKPROPAGATION;
D O I
10.3389/fnins.2021.756876
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Spiking Neural Networks (SNNs) are a pathway that could potentially empower low-power event-driven neuromorphic hardware due to their spatio-temporal information processing capability and high biological plausibility. Although SNNs are currently more efficient than artificial neural networks (ANNs), they are not as accurate as ANNs. Error backpropagation is the most common method for directly training neural networks, promoting the prosperity of ANNs in various deep learning fields. However, since the signals transmitted in the SNN are non-differentiable discrete binary spike events, the activation function in the form of spikes presents difficulties for the gradient-based optimization algorithms to be directly applied in SNNs, leading to a performance gap (i.e., accuracy and latency) between SNNs and ANNs. This paper introduces a new learning algorithm, called SSTDP, which bridges the gap between backpropagation (BP)-based learning and spike-time-dependent plasticity (STDP)-based learning to train SNNs efficiently. The scheme incorporates the global optimization process from BP and the efficient weight update derived from STDP. It not only avoids the non-differentiable derivation in the BP process but also utilizes the local feature extraction property of STDP. Consequently, our method can lower the possibility of vanishing spikes in BP training and reduce the number of time steps to reduce network latency. In SSTDP, we employ temporal-based coding and use Integrate-and-Fire (IF) neuron as the neuron model to provide considerable computational benefits. Our experiments show the effectiveness of the proposed SSTDP learning algorithm on the SNN by achieving the best classification accuracy 99.3% on the Caltech 101 dataset, 98.1% on the MNIST dataset, and 91.3% on the CIFAR-10 dataset compared to other SNNs trained with other learning methods. It also surpasses the best inference accuracy of the directly trained SNN with 25 similar to 32x less inference latency. Moreover, we analyze event-based computations to demonstrate the efficacy of the SNN for inference operation in the spiking domain, and SSTDP methods can achieve 1.3 similar to 37.7x fewer addition operations per inference. The code is available at: https://github.com/MXHX7199/SNN-SSTDP.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] True North: Design and Tool Flow of a 65 mW 1 Million Neuron Programmable Neurosynaptic Chip
    Akopyan, Filipp
    Sawada, Jun
    Cassidy, Andrew
    Alvarez-Icaza, Rodrigo
    Arthur, John
    Merolla, Paul
    Imam, Nabil
    Nakamura, Yutaka
    Datta, Pallab
    Nam, Gi-Joon
    Taba, Brian
    Beakes, Michael
    Brezzo, Bernard
    Kuang, Jente B.
    Manohar, Rajit
    Risk, William P.
    Jackson, Bryan
    Modha, Dharmendra S.
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2015, 34 (10) : 1537 - 1557
  • [2] Bellec G, 2018, 32 C NEURAL INFORM P
  • [3] Neurogrid: A Mixed-Analog-Digital Multichip System for Large-Scale Neural Simulations
    Benjamin, Ben Varkey
    Gao, Peiran
    McQuinn, Emmett
    Choudhary, Swadesh
    Chandrasekaran, Anand R.
    Bussat, Jean-Marie
    Alvarez-Icaza, Rodrigo
    Arthur, John V.
    Merolla, Paul A.
    Boahen, Kwabena
    [J]. PROCEEDINGS OF THE IEEE, 2014, 102 (05) : 699 - 716
  • [4] Bing Han, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12355), P388, DOI 10.1007/978-3-030-58607-2_23
  • [5] Bing Han, 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Proceedings, P13555, DOI 10.1109/CVPR42600.2020.01357
  • [6] Comsa JM, 2020, INT CONF ACOUST SPEE, P8529, DOI [10.1109/icassp40776.2020.9053856, 10.1109/ICASSP40776.2020.9053856]
  • [7] Loihi: A Neuromorphic Manycore Processor with On-Chip Learning
    Davies, Mike
    Srinivasa, Narayan
    Lin, Tsung-Han
    Chinya, Gautham
    Cao, Yongqiang
    Choday, Sri Harsha
    Dimou, Georgios
    Joshi, Prasad
    Imam, Nabil
    Jain, Shweta
    Liao, Yuyun
    Lin, Chit-Kwan
    Lines, Andrew
    Liu, Ruokun
    Mathaikutty, Deepak
    Mccoy, Steve
    Paul, Arnab
    Tse, Jonathan
    Venkataramanan, Guruguhanathan
    Weng, Yi-Hsin
    Wild, Andreas
    Yang, Yoonseok
    Wang, Hong
    [J]. IEEE MICRO, 2018, 38 (01) : 82 - 99
  • [8] Editorial: Understanding and Bridging the Gap Between Neuromorphic Computing and Machine Learning
    Deng, Lei
    Tang, Huajin
    Roy, Kaushik
    [J]. FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2021, 15
  • [9] Rethinking the performance comparison between SNNS and ANNS
    Deng, Lei
    Wu, Yujie
    Hu, Xing
    Liang, Ling
    Ding, Yufei
    Li, Guoqi
    Zhao, Guangshe
    Li, Peng
    Xie, Yuan
    [J]. NEURAL NETWORKS, 2020, 121 : 294 - 307
  • [10] Devlin J, 2019, Arxiv, DOI [arXiv:1810.04805, DOI 10.48550/ARXIV.1810.04805]