Effective Surrogate Gradient Learning With High-Order Information Bottleneck for Spike-Based Machine Intelligence

被引:43
作者
Yang, Shuangming [1 ]
Chen, Badong [2 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin 300072, Peoples R China
[2] Xi An Jiao Tong Univ, Inst Artificial Intelligence & Robot, Xian 710049, Peoples R China
基金
中国国家自然科学基金;
关键词
Training; Neurons; Robustness; Mutual information; Biological neural networks; Power demand; Kernel; Information bottleneck (IB); information-theoretic learning (ITL); neuromorphic computing; spike-driven learning; spiking neural network (SNN); NEURAL-NETWORKS;
D O I
10.1109/TNNLS.2023.3329525
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Brain-inspired computing technique presents a promising approach to prompt the rapid development of artificial general intelligence (AGI). As one of the most critical aspects, spiking neural networks (SNNs) have demonstrated superiority for AGI, such as low power consumption. Effective training of SNNs with high generalization ability, high robustness, and low power consumption simultaneously is a significantly challenging problem for the development and success of applications of spike-based machine intelligence. In this research, we present a novel and flexible learning framework termed high-order spike-based information bottleneck (HOSIB) leveraging the surrogate gradient technique. The presented HOSIB framework, including second-order and third-order formation, i.e., second-order information bottleneck (SOIB) and third-order information bottleneck (TOIB), comprehensively explores the common latent architecture and the spike-based intrinsic information and discards the superfluous information in the data, which improves the generalization capability and robustness of SNN models. Specifically, HOSIB relies on the information bottleneck (IB) principle to prompt the sparse spike-based information representation and flexibly balance its exploitation and loss. Extensive classification experiments are conducted to empirically show the promising generalization ability of HOSIB. Furthermore, we apply the SOIB and TOIB algorithms in deep spiking convolutional networks to demonstrate their improvement in robustness with various categories of noise. The experimental results prove the HOSIB framework, especially TOIB, can achieve better generalization ability, robustness and power efficiency in comparison with the current representative studies.
引用
收藏
页码:1734 / 1748
页数:15
相关论文
共 64 条
[21]  
Kheradpisheh SR, 2020, Arxiv, DOI arXiv:1910.09495
[22]   Revisiting Batch Normalization for Training Low-Latency Deep Spiking Neural Networks From Scratch [J].
Kim, Youngeun ;
Panda, Priyadarshini .
FRONTIERS IN NEUROSCIENCE, 2021, 15
[23]   Deep learning [J].
LeCun, Yann ;
Bengio, Yoshua ;
Hinton, Geoffrey .
NATURE, 2015, 521 (7553) :436-444
[24]   Training Deep Spiking Neural Networks Using Backpropagation [J].
Lee, Jun Haeng ;
Delbruck, Tobi ;
Pfeiffer, Michael .
FRONTIERS IN NEUROSCIENCE, 2016, 10
[25]   Deep Hybrid 2-D-3-D CNN Based on Dual Second-Order Attention With Camera Spectral Sensitivity Prior for Spectral Super-Resolution [J].
Li, Jiaojiao ;
Wu, Chaoxiong ;
Song, Rui ;
Li, Yunsong ;
Xie, Weiying ;
He, Lihuo ;
Gao, Xinbo .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (02) :623-634
[26]  
Liu J., 2020, Simple and principled uncertainty estimation with deterministic deep learning via distance awareness, V33, P7498
[27]   Training High-Performance Low-Latency Spiking Neural Networks by Differentiation on Spike Representation [J].
Meng, Qingyan ;
Xiao, Mingqing ;
Yan, Shen ;
Wang, Yisen ;
Lin, Zhouchen ;
Luo, Zhi-Quan .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :12434-12443
[28]   Surrogate Gradient Learning in Spiking Neural Networks: Bringing the Power of Gradient-based optimization to spiking neural networks [J].
Neftci, Emre O. ;
Mostafa, Hesham ;
Zenke, Friedemann .
IEEE SIGNAL PROCESSING MAGAZINE, 2019, 36 (06) :51-63
[29]   A novel and efficient classifier using spiking neural network [J].
Ranjan, Joshua Arul Kumar ;
Sigamani, Titus ;
Barnabas, Janet .
JOURNAL OF SUPERCOMPUTING, 2020, 76 (09) :6545-6560
[30]  
Rathi N, 2020, Arxiv, DOI arXiv:2005.01807