Training Spiking Neural Networks with Local Tandem Learning

被引:0
作者
Yang, Qu [1 ]
Wu, Jibin [2 ]
Zhang, Malu [3 ]
Chua, Yansong [4 ]
Wang, Xinchao [1 ]
Li, Haizhou [1 ,5 ,6 ]
机构
[1] Natl Univ Singapore, Singapore, Singapore
[2] Hong Kong Polytech Univ, Hong Kong, Peoples R China
[3] Univ Elect Sci & Technol China, Chengdu, Peoples R China
[4] China Nanhu Acad Elect & Informat Technol, Beijing, Peoples R China
[5] Chinese Univ Hong Kong, Shenzhen, Peoples R China
[6] Kriston AI, Xiamen, Peoples R China
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022) | 2022年
基金
美国国家科学基金会; 新加坡国家研究基金会;
关键词
INTELLIGENCE; POWER;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Spiking neural networks (SNNs) are shown to be more biologically plausible and energy efficient over their predecessors. However, there is a lack of an efficient and generalized training method for deep SNNs, especially for deployment on analog computing substrates. In this paper, we put forward a generalized learning rule, termed Local Tandem Learning (LTL). The LTL rule follows the teacher-student learning approach by mimicking the intermediate feature representations of a pre-trained ANN. By decoupling the learning of network layers and leveraging highly informative supervisor signals, we demonstrate rapid network convergence within five training epochs on the CIFAR-10 dataset while having low computational complexity. Our experimental results have also shown that the SNNs thus trained can achieve comparable accuracies to their teacher ANNs on CIFAR-10, CIFAR-100, and Tiny ImageNet datasets. Moreover, the proposed LTL rule is hardware friendly. It can be easily implemented on-chip to perform fast parameter calibration and provide robustness against the notorious device non-ideality issues. It, therefore, opens up a myriad of opportunities for training and deployment of SNN on ultra-low-power mixed-signal neuromorphic computing chips.
引用
收藏
页数:15
相关论文
共 64 条
[1]   An Accelerated LIF Neuronal Network Array for a Large-Scale Mixed-Signal Neuromorphic Architecture [J].
Aamir, Syed Ahmed ;
Stradmann, Yannik ;
Mueller, Paul ;
Pehle, Christian ;
Hartel, Andreas ;
Gruebl, Andreas ;
Schemmel, Johannes ;
Meier, Karlheinz .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2018, 65 (12) :4299-4312
[2]   True North: Design and Tool Flow of a 65 mW 1 Million Neuron Programmable Neurosynaptic Chip [J].
Akopyan, Filipp ;
Sawada, Jun ;
Cassidy, Andrew ;
Alvarez-Icaza, Rodrigo ;
Arthur, John ;
Merolla, Paul ;
Imam, Nabil ;
Nakamura, Yutaka ;
Datta, Pallab ;
Nam, Gi-Joon ;
Taba, Brian ;
Beakes, Michael ;
Brezzo, Bernard ;
Kuang, Jente B. ;
Manohar, Rajit ;
Risk, William P. ;
Jackson, Bryan ;
Modha, Dharmendra S. .
IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2015, 34 (10) :1537-1557
[3]  
Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
[4]  
Bajracharya A, 2018, PROCEEDINGS OF THE 8TH INTERNATIONAL CONFERENCE CONFLUENCE 2018 ON CLOUD COMPUTING, DATA SCIENCE AND ENGINEERING, P314, DOI 10.1109/CONFLUENCE.2018.8442732
[5]  
Bu T, 2021, INT C LEARN REPR
[6]  
Bu Tong, 2022, ARXIV220201440
[7]   Supervised training of spiking neural networks for robust deployment on mixed-signal neuromorphic processors [J].
Buechel, Julian ;
Zendrikov, Dmitrii ;
Solinas, Sergio ;
Indiveri, Giacomo ;
Muir, Dylan R. .
SCIENTIFIC REPORTS, 2021, 11 (01)
[8]   Spiking Deep Convolutional Neural Networks for Energy-Efficient Object Recognition [J].
Cao, Yongqiang ;
Chen, Yang ;
Khosla, Deepak .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2015, 113 (01) :54-66
[9]  
Chen Yanqi, 2021, ARXIV210504916
[10]   Analog Memristive Synapse in Spiking Networks Implementing Unsupervised Learning [J].
Covi, Erika ;
Brivio, Stefano ;
Serb, Alexander ;
Prodromakis, Themis ;
Fanciulli, Marco ;
Spiga, Sabina .
FRONTIERS IN NEUROSCIENCE, 2016, 10