Exploiting memristive autapse and temporal distillation for training spiking neural networks

被引:0
作者
Chen, Tao [1 ]
Duan, Shukai [1 ,2 ,3 ]
Wang, Lidan [1 ,2 ,3 ,4 ,5 ]
机构
[1] Southwest Univ, Coll Artificial Intelligence, Chongqing 400715, Peoples R China
[2] Chongqing Key Lab Brain inspired Comp & Intelligen, Chongqing 400715, Peoples R China
[3] Natl & Local Joint Engn Lab Intelligent Transmiss, Chongqing 400715, Peoples R China
[4] Southwest Univ, Minist Educ, Key Lab Luminescence Anal & Mol Sensing, Chongqing 400715, Peoples R China
[5] State Key Lab Intelligent Vehicle Safety Technol, Chongqing 400023, Peoples R China
基金
中国国家自然科学基金;
关键词
Spiking neural network; Memristive autapse; Adaptive self-feedback connections; Temporal information; Knowledge distillation; NEURONS;
D O I
10.1016/j.knosys.2024.112627
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Spiking neural networks (SNNs) have attracted widespread attention due to their brain-inspired information processing mechanism and low power, sparse accumulation computation on neuromorphic chips. The surrogate gradient method makes it possible to train deep SNNs using backpropagation and shows satisfactory performance on some tasks. However, as the network structure becomes deeper, the spike information may fail to transmit to deeper layers, thus causing the output layer to make wrong predictions in recognition tasks. Inspired by the autaptic structure in the cerebral cortex, which is formed by axons connecting to their own dendrites and capable of modulating neuronal activity, we use discrete memristors to build feedback-connected autapses to adaptively regulate the precision of the spikes. Further, to prevent outlier at a certain time step from affecting the overall output, we distill the averaged knowledge into sub-models at each time step to correct potential errors. By combining these two proposed methods, we propose a deep SNNs optimized by Leaky Integrate-and-Fire (LIF) model with memristive autapse and temporal distillation, referred to as MASNN. A series of experiments on static datasets (CIFAR10 and CIFAR100) as well as neuromorphic datasets (DVS-CIFAR10 and N-Caltech101) demonstrated the competitiveness of the proposed model and validated the effectiveness of its components. Code for MA-SNN is available at: https://github.com/CHNtao/MA-SNN.
引用
收藏
页数:12
相关论文
共 53 条
[1]   Hidden Bursting Firings and Bifurcation Mechanisms in Memristive Neuron Model With Threshold Electromagnetic Induction [J].
Bao, Han ;
Hu, Aihuang ;
Liu, Wenbo ;
Bao, Bocheng .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (02) :502-511
[2]   Spiking Deep Convolutional Neural Networks for Energy-Efficient Object Recognition [J].
Cao, Yongqiang ;
Chen, Yang ;
Khosla, Deepak .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2015, 113 (01) :54-66
[3]  
Chen T., 2023, IEEE Transactions on Cognitive and Developmental Systems
[4]   Memristive leaky integrate-and-fire neuron and learnable straight-through estimator in spiking neural networks [J].
Chen, Tao ;
She, Chunyan ;
Wang, Lidan ;
Duan, Shukai .
COGNITIVE NEURODYNAMICS, 2024, 18 (05) :3075-3091
[5]   Surrogate gradient scaling for directly training spiking neural networks [J].
Chen, Tao ;
Wang, Shu ;
Gong, Yu ;
Wang, Lidan ;
Duan, Shukai .
APPLIED INTELLIGENCE, 2023, 53 (23) :27966-27981
[6]   Loihi: A Neuromorphic Manycore Processor with On-Chip Learning [J].
Davies, Mike ;
Srinivasa, Narayan ;
Lin, Tsung-Han ;
Chinya, Gautham ;
Cao, Yongqiang ;
Choday, Sri Harsha ;
Dimou, Georgios ;
Joshi, Prasad ;
Imam, Nabil ;
Jain, Shweta ;
Liao, Yuyun ;
Lin, Chit-Kwan ;
Lines, Andrew ;
Liu, Ruokun ;
Mathaikutty, Deepak ;
Mccoy, Steve ;
Paul, Arnab ;
Tse, Jonathan ;
Venkataramanan, Guruguhanathan ;
Weng, Yi-Hsin ;
Wild, Andreas ;
Yang, Yoonseok ;
Wang, Hong .
IEEE MICRO, 2018, 38 (01) :82-99
[7]   TrueNorth: Accelerating From Zero to 64 Million Neurons in 10 Years [J].
DeBole, Michael V. ;
Taba, Brian ;
Amir, Arnon ;
Akopyan, Filipp ;
Andreopoulos, Alexander ;
Risk, William P. ;
Kusnitz, Jeff ;
Otero, Carlos Ortega ;
Nayak, Tapan K. ;
Appuswamy, Rathinakumar ;
Carlson, Peter J. ;
Cassidy, Andrew S. ;
Datta, Pallab ;
Esser, Steven K. ;
Garreau, Guillaume J. ;
Holland, Kevin L. ;
Lekuch, Scott ;
Mastro, Michael ;
McKinstry, Jeff ;
di Nolfo, Carmelo ;
Paulovicks, Brent ;
Sawada, Jun ;
Schleupen, Kai ;
Shaw, Benjamin G. ;
Klamo, Jennifer L. ;
Flickner, Myron D. ;
Arthur, John V. ;
Modha, Dharmendra S. .
COMPUTER, 2019, 52 (05) :20-29
[8]  
Deng S., 2022, INT C LEARN REPR
[9]   An improved probabilistic spiking neural network with enhanced discriminative ability [J].
Ding, Yongqi ;
Zuo, Lin ;
Yang, Kunshan ;
Chen, Zhongshu ;
Hu, Jian ;
Xiahou, Tangfan .
KNOWLEDGE-BASED SYSTEMS, 2023, 280
[10]  
Dong Z., 2023, IEEE Trans. Consum. Electron.