Surrogate gradient scaling for directly training spiking neural networks

被引:0
作者
Tao Chen
Shu Wang
Yu Gong
Lidan Wang
Shukai Duan
机构
[1] Southwest University,College of Artificial Intelligence
[2] Brain-Inspired Computing & Intelligent Control of Chongqing Key Lab,undefined
[3] National & Local Joint Engineering Laboratory of Intelligent Transmission and Control Technology,undefined
[4] Chongqing Brain Science Collaborative Innovation Center,undefined
来源
Applied Intelligence | 2023年 / 53卷
关键词
Neuromorphic computing; Lipschitz regularization; Energy efficient; Noisy robust; Neural coding; Image recognition;
D O I
暂无
中图分类号
学科分类号
摘要
Spiking neural networks (SNNs) are considered to be biologically plausible and can yield high energy efficiency when implemented on neuromorphic hardware due to their highly sparse asynchronous binary event-driven nature. Recently, surrogate gradient (SG) approaches have enabled SNNs to be trained from scratch with backpropagation (BP) algorithms under a deep learning framework. However, a popular SG approach known as straight-through estimator (STE), which only propagates the same gradient information, does not take into account the activation differences between the membrane potentials and output spikes. To address this issue, we propose surrogate gradient scaling (SGS), which scales up or down the gradient information of the membrane potential according to the sign of the gradient of the spiking neuron output and the difference between the membrane potential and the output of the spiking neuron. This SGS approach can also be applied to unimodal functions that propagate different gradient information from the output spikes to the input membrane potential. In addition, SNNs trained directly from scratch suffer from poor generalization performance, and we introduce Lipschitz regularization (LR), which is incorporated into the loss function. It not only improves the generalization performance of SNNs but also makes them more robust to noise. Extensive experimental results on several popular benchmark datasets (CIFAR10, CIFAR100 and CIFAR10-DVS) show that our approach not only outperforms the SOTA but also has lower inference latency. Remarkably, our SNNs can lead to 34×\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}, 29×\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document}, and 17×\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times $$\end{document} computation energy savings compared to standard Artificial neural networks (ANNs) on above three datasets.
引用
收藏
页码:27966 / 27981
页数:15
相关论文
共 31 条
[1]  
Zhao ZQ(2019)Object detection with deep learning: A review IEEE Trans Neural Netw Learn Syst 30 3212-3232
[2]  
Zheng P(2019)Towards spike-based machine intelligence with neuromorphic computing Nature 575 607-617
[3]  
St Xu(1997)Networks of spiking neurons: the third generation of neural network models Neural Netw, 10 1659-1671
[4]  
Roy K(2019)Sparse computation in adaptive spiking neural networks Front Neurosci, 12 987-99
[5]  
Jaiswal A(2020)Toward scalable, efficient, and accurate deep spiking neural networks with backward residual connections, stochastic softmax, and hybridization Front Neurosci, 14 653-111
[6]  
Panda P(2018)Loihi: A neuromorphic manycore processor with on-chip learning IEEE Micro, 38 82-510
[7]  
Maass W(2019)Towards artificial general intelligence with hybrid tianjic chip architecture Nature 572 106-428
[8]  
Zambrano D(2010)Supervised learning in spiking neural networks with resume: sequence learning, classification, and spike shifting Neural Comput, 22 467-444
[9]  
Nusselder R(2006)The tempotron: A neuron that learns spike timing-based decisions Nat Neurosci, 9 420-63
[10]  
Scholte HS(2015)Deep learning Nature 521 436-1541