IM-LIF: Improved Neuronal Dynamics With Attention Mechanism for Direct Training Deep Spiking Neural Network

被引:4
作者
Lian, Shuang [1 ]
Shen, Jiangrong [1 ]
Wang, Ziming [1 ]
Tang, Huajin [1 ,2 ]
机构
[1] Zhejiang Univ City Coll, Coll Comp Sci & Technol, Hangzhou 310027, Peoples R China
[2] Zhejiang Lab, Hangzhou 311122, Peoples R China
来源
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE | 2024年 / 8卷 / 02期
基金
中国国家自然科学基金;
关键词
Spiking neural networks; attention mechanism; spiking neuron; gradient vanishing; backpropagation;
D O I
10.1109/TETCI.2024.3359539
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Spiking neural networks (SNNs) are increasingly applied to deep architectures. Recent works are developed to apply spatio-temporal backpropagation to directly train deep SNNs. But the binary and non-differentiable properties of spike activities force directly trained SNNs to suffer from serious gradient vanishing. In this paper, we first analyze the cause of the gradient vanishing problem and identify that the gradients mostly backpropagate along the synaptic currents. Based on that, we modify the synaptic current equation of leaky-integrate-fire neuron model and propose the improved LIF (IM-LIF) neuron model on the basis of the temporal-wise attention mechanism. We utilize the temporal-wise attention mechanism to selectively establish the connection between the current and historical response values, which can empirically enable the neuronal states to update resilient to the gradient vanishing problem. Furthermore, to capture the neuronal dynamics embedded in the output incorporating the IM-LIF model, we present a new temporal loss function to constrain the output of the network close to the target distribution. The proposed new temporal loss function could not only act as a regularizer to eliminate output outliers, but also assign the network loss credit to the voltage at a specific time point. Then we modify the ResNet and VGG architecture based on the IM-LIF model to build deep SNNs. We evaluate our work on image datasets and neuromorphic datasets. Experimental results and analysis show that our method can help build deep SNNs with competitive performance in both accuracy and latency, including 95.66% on CIFAR-10, 77.42% on CIFAR-100, 55.37% on Tiny-ImageNet, 97.33% on DVS-Gesture, and 80.50% on CIFAR-DVS with very few timesteps.
引用
收藏
页码:2075 / 2085
页数:11
相关论文
共 61 条
  • [1] Amir Arnon, 2017, P IEEE C COMP VIS PA, P7243, DOI DOI 10.1109/CVPR.2017.781
  • [2] Bu T., 2022, P 10 INT C LEARN REP
  • [3] Attention Mechanisms for Object Recognition with Event-Based Cameras
    Cannici, Marco
    Ciccone, Marco
    Romanoni, Andrea
    Matteucci, Matteo
    [J]. 2019 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2019, : 1127 - 1136
  • [4] Cheng X, 2020, PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P1519
  • [5] Chung J., 2014, arXiv
  • [6] Deng S., 2022, P INT C LEARN REPR, P1
  • [7] Duan CT, 2022, ADV NEUR IN
  • [8] Fang W, 2021, ADV NEUR IN, V34
  • [9] Incorporating Learnable Membrane Time Constant to Enhance Learning of Spiking Neural Networks
    Fang, Wei
    Yu, Zhaofei
    Chen, Yanqi
    Masquelier, Timothee
    Huang, Tiejun
    Tian, Yonghong
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 2641 - 2651
  • [10] Feng L, 2022, PROCEEDINGS OF THE THIRTY-FIRST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2022, P2471