LDD: High-Precision Training of Deep Spiking Neural Network Transformers Guided by an Artificial Neural Network

被引:1
|
作者
Liu, Yuqian [1 ,2 ]
Zhao, Chujie [1 ,2 ]
Jiang, Yizhou [1 ,2 ]
Fang, Ying [3 ,4 ]
Chen, Feng [1 ,2 ]
机构
[1] Tsinghua Univ, Dept Automat, Beijing 100084, Peoples R China
[2] LSBDPA Beijing Key Lab, Beijing 100084, Peoples R China
[3] Fujian Normal Univ, Coll Comp & Cyber Secur, Fuzhou 350117, Peoples R China
[4] Fujian Normal Univ, Digital Fujian Internet of Thing Lab Environm Moni, Fuzhou 350117, Peoples R China
基金
中国国家自然科学基金;
关键词
spiking neural networks (SNNs); Transformer; distillation; image classification;
D O I
10.3390/biomimetics9070413
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
The rise of large-scale Transformers has led to challenges regarding computational costs and energy consumption. In this context, spiking neural networks (SNNs) offer potential solutions due to their energy efficiency and processing speed. However, the inaccuracy of surrogate gradients and feature space quantization pose challenges for directly training deep SNN Transformers. To tackle these challenges, we propose a method (called LDD) to align ANN and SNN features across different abstraction levels in a Transformer network. LDD incorporates structured feature knowledge from ANNs to guide SNN training, ensuring the preservation of crucial information and addressing inaccuracies in surrogate gradients through designing layer-wise distillation losses. The proposed approach outperforms existing methods on the CIFAR10 (96.1%), CIFAR100 (82.3%), and ImageNet (80.9%) datasets, and enables training of the deepest SNN Transformer network using ImageNet.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Parallel proportional fusion of a spiking quantum neural network for optimizing image classification
    Xu, Zuyu
    Shen, Kang
    Cai, Pengnian
    Yang, Tao
    Hu, Yuanming
    Chen, Shixian
    Zhu, Yunlai
    Wu, Zuheng
    Dai, Yuehua
    Wang, Jun
    Yang, Fei
    APPLIED INTELLIGENCE, 2024, 54 (22) : 11876 - 11891
  • [32] PCFNet: Deep neural network with predefined convolutional filters
    Ma, Yangling
    Luo, Yixin
    Yang, Zhouwang
    NEUROCOMPUTING, 2020, 382 (382) : 32 - 39
  • [33] ErfReLU: adaptive activation function for deep neural network
    Rajanand, Ashish
    Singh, Pradeep
    PATTERN ANALYSIS AND APPLICATIONS, 2024, 27 (02)
  • [34] Deep Convolutional Neural Network for Chicken Diseases Detection
    Mbelwa, Hope
    Machuve, Dina
    Mbelwa, Jimmy
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2021, 12 (02) : 759 - 765
  • [35] Variational quantum deep neural network for image classification
    Xu, Fangling
    Zhang, Xuesong
    SCIENTIA SINICA-PHYSICA MECHANICA & ASTRONOMICA, 2025, 55 (03)
  • [36] TE-Spikformer:Temporal-enhanced spiking neural network with transformer
    Gao, ShouWei
    Fan, XiangYu
    Deng, XingYang
    Hong, ZiChao
    Zhou, Hao
    Zhu, ZiHao
    NEUROCOMPUTING, 2024, 602
  • [37] Low Power Spiking Neural Network Circuit with Compact Synapse and Neuron Cells
    Asghar, Malik Summair
    Arslan, Saad
    Kim, HyungWon
    2020 17TH INTERNATIONAL SOC DESIGN CONFERENCE (ISOCC 2020), 2020, : 157 - 158
  • [38] Study of the influence of synaptic plasticity on the formation of a feature spaceby a spiking neural network
    Lebedev, A. A.
    Kazantsev, V. B.
    V. Stasenko, S.
    IZVESTIYA VYSSHIKH UCHEBNYKH ZAVEDENIY-PRIKLADNAYA NELINEYNAYA DINAMIKA, 2024, 32 (02): : 253 - 267
  • [39] PatternNet: Visual Pattern Mining with Deep Neural Network
    Li, Hongzhi
    Ellis, Joseph G.
    Zhang, Lei
    Chang, Shih-Fu
    ICMR '18: PROCEEDINGS OF THE 2018 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, 2018, : 291 - 299
  • [40] A high-precision image classification network model based on a voting mechanism
    Zhao, Jianghong
    Wang, Xin
    Dou, Xintong
    Zhao, Yingxue
    Fu, Zexin
    Guo, Ming
    Zhang, Ruiju
    INTERNATIONAL JOURNAL OF DIGITAL EARTH, 2022, 15 (01) : 2168 - 2183