Knowledge Distill for Spiking Neural Networks

被引:0
作者
Liang, Xiubo [1 ]
Chao, Ge [1 ]
Li, Mengjian [2 ]
Zhao, Yijun [1 ]
机构
[1] Zhejiang Univ, Sch Software Technol, Ningbo, Peoples R China
[2] Zhejiang Lab, Res Ctr Data Hub & Secur, Hangzhou, Peoples R China
来源
2024 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN 2024 | 2024年
关键词
spiking neural network; knowledge-distill; INTELLIGENCE; NEURONS;
D O I
10.1109/IJCNN60899.2024.10650960
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Spiking Neural Network (SNN) is a kind of braininspired and event-driven network, which is becoming a promising energy-efficient alternative to Artificial Neural Networks (ANNs). However, the performance of SNNs by direct training is far from satisfactory. Inspired by the idea of Teacher-Student Learning, in this paper, we study a novel learning method named SuperSNN, which utilizes the ANN model to guide the SNN model learning. SuperSNN leverages knowledge distillation to learn comprehensive supervisory information from pre-trained ANN models, rather than solely from labeled data. Unlike previous work that naively matches SNN and ANN's features without deeply considering the precision mismatch, we propose an indirect relation-based approach, which defines a pairwiserelational loss function and unifies the value scale of ANN and SNN representation vectors, to alleviate the unexpected precision loss. This allows the knowledge of teacher ANNs can be effectively utilized to train student SNNs. The experimental results on three image datasets demonstrate that no matter whether homogeneous or heterogeneous teacher ANNs are used, our proposed SuperSNN can significantly improve the learning of student SNNs with only two time steps.
引用
收藏
页数:8
相关论文
共 57 条
[1]   Variational Information Distillation for Knowledge Transfer [J].
Ahn, Sungsoo ;
Hu, Shell Xu ;
Damianou, Andreas ;
Lawrence, Neil D. ;
Dai, Zhenwen .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9155-9163
[2]  
[Anonymous], 1989, The Technical Writer's Handbook
[3]  
[Anonymous], 2019, Surrogate gradient learning in spiking neural networks
[4]  
[Anonymous], 2016, DEEP PYRAMIDAL RESID
[5]  
Ba J., 2014, C NEUR INF PROC SYST, V27, P2654
[6]  
Baig Z., 2015, COMPUTING REV, V56, P747
[7]  
Brutzkus A., 2018, DO LARGER MODELS GEN
[8]  
Bucila C., 2006, ACM SIGKDD INT C KNO
[9]   DSSA: Dual-Side Sparse Systolic Array Architecture for Accelerating Convolutional Neural Network Training [J].
Chen, Zhengbo ;
Yu, Qi ;
Zheng, Fang ;
Guo, Feng ;
Chen, Zuoning .
51ST INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, ICPP 2022, 2022,
[10]  
Clerk Maxwell J., 1892, A Treatise on Electricity and Magnetism, V2, P68