SPIDE: A purely spike-based method for training feedback spiking neural networks

被引:8
作者
Xiao, Mingqing [1 ]
Meng, Qingyan [2 ,3 ]
Zhang, Zongpeng [4 ]
Wang, Yisen [1 ,5 ]
Lin, Zhouchen [1 ,5 ,6 ]
机构
[1] Peking Univ, Sch Intelligence Sci & Technol, Natl Key Lab Gen Artificial Intelligence, Beijing, Peoples R China
[2] Chinese Univ Hong Kong, Shenzhen, Peoples R China
[3] Shenzhen Res Inst Big Data, Shenzhen 518115, Peoples R China
[4] Peking Univ, Acad Adv Interdisciplinary Studies, Ctr Data Sci, Beijing, Peoples R China
[5] Peking Univ, Inst Artificial Intelligence, Beijing, Peoples R China
[6] Peng Cheng Lab, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
Spiking neural networks; Equilibrium state; Spike-based training method; Neuromorphic computing; BACKPROPAGATION; NEURONS; INTELLIGENCE;
D O I
10.1016/j.neunet.2023.01.026
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Spiking neural networks (SNNs) with event-based computation are promising brain-inspired models for energy-efficient applications on neuromorphic hardware. However, most supervised SNN training methods, such as conversion from artificial neural networks or direct training with surrogate gradients, require complex computation rather than spike-based operations of spiking neurons during training. In this paper, we study spike-based implicit differentiation on the equilibrium state (SPIDE) that extends the recently proposed training method, implicit differentiation on the equilibrium state (IDE), for supervised learning with purely spike-based computation, which demonstrates the potential for energy-efficient training of SNNs. Specifically, we introduce ternary spiking neuron couples and prove that implicit differentiation can be solved by spikes based on this design, so the whole training procedure, including both forward and backward passes, is made as event-driven spike computation, and weights are updated locally with two-stage average firing rates. Then we propose to modify the reset membrane potential to reduce the approximation error of spikes. With these key components, we can train SNNs with flexible structures in a small number of time steps and with firing sparsity during training, and the theoretical estimation of energy costs demonstrates the potential for high efficiency. Meanwhile, experiments show that even with these constraints, our trained models can still achieve competitive results on MNIST, CIFAR-10, CIFAR-100, and CIFAR10-DVS.(c) 2023 Published by Elsevier Ltd.
引用
收藏
页码:9 / 24
页数:16
相关论文
共 78 条
  • [1] True North: Design and Tool Flow of a 65 mW 1 Million Neuron Programmable Neurosynaptic Chip
    Akopyan, Filipp
    Sawada, Jun
    Cassidy, Andrew
    Alvarez-Icaza, Rodrigo
    Arthur, John
    Merolla, Paul
    Imam, Nabil
    Nakamura, Yutaka
    Datta, Pallab
    Nam, Gi-Joon
    Taba, Brian
    Beakes, Michael
    Brezzo, Bernard
    Kuang, Jente B.
    Manohar, Rajit
    Risk, William P.
    Jackson, Bryan
    Modha, Dharmendra S.
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2015, 34 (10) : 1537 - 1557
  • [2] Almeida L., 1987, INT C NEUR NETW
  • [3] [Anonymous], 2018, Advances in Neural Information Processing Systems
  • [4] [Anonymous], 2009, Rep. TR-2009
  • [5] Bai S., 2020, Advances in Neural Information Processing Systems (NeurIPS)
  • [6] Bai Shaojie, 2021, INT C MACHINE LEARNI
  • [7] Bai Shaojie, 2019, Neural Information Processing Systems, V1
  • [8] Bellec G, 2018, 32 C NEURAL INFORM P
  • [9] Bichler O., 2019, IEEE IJCNN
  • [10] Error-backpropagation in temporally encoded networks of spiking neurons
    Bohte, SM
    Kok, JN
    La Poutré, H
    [J]. NEUROCOMPUTING, 2002, 48 : 17 - 37