Tensor decomposition based attention module for spiking neural networks

被引:3
作者
Deng, Haoyu [1 ]
Zhu, Ruijie [2 ]
Qiu, Xuerui [1 ]
Duan, Yule [1 ]
Zhang, Malu [1 ]
Deng, Liang-Jian [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu 611731, Peoples R China
[2] Univ Calif Santa Cruz, Santa Cruz, CA 95064 USA
基金
美国国家科学基金会;
关键词
Spiking neural network; Attention mechanism; Tensor decomposition; Neuromorphic computing; DEEPER;
D O I
10.1016/j.knosys.2024.111780
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The attention mechanism has been proven to be an effective way to improve the performance of spiking neural networks (SNNs). However, from the perspective of tensor decomposition to examine the existing attention modules, we find that the rank of the attention maps generated by previous methods is fixed at 1, lacking the flexibility to adjust for specific tasks. To tackle this problem, we propose an attention module, namely Projected-full Attention (PFA), where the rank of the generated attention maps can be determined based on the characteristics of different tasks. Additionally, the parameter count of PFA grows linearly with the data scale. PFA is composed of the linear projection of spike tensor (LPST) module and attention map composing (AMC) module. In LPST, we start by compressing the original spike tensor into three projected tensors with learnable parameters for each dimension. Then, in AMC, we exploit the inverse procedure of the tensor decomposition process to combine the three tensors into the attention map using a so-called connecting factor. To validate the effectiveness of the proposed PFA module, we integrate it into the widely used VGG and ResNet architectures for classification tasks. Our method achieves state-of-the-art performance on both static and dynamic benchmark datasets, surpassing the existing SNN models with Transformer-based and CNN-based backbones. Code for PFA is available at https://github.com/RisingEntropy/PFA.
引用
收藏
页数:10
相关论文
共 66 条
[1]   Scalable tensor factorizations for incomplete data [J].
Acar, Evrim ;
Dunlavy, Daniel M. ;
Kolda, Tamara G. ;
Morup, Morten .
CHEMOMETRICS AND INTELLIGENT LABORATORY SYSTEMS, 2011, 106 (01) :41-56
[2]  
[Anonymous], 2012, IEEE Signal Process. Mag., DOI DOI 10.1109/MSP.2012.2211477
[3]   Factorized Dynamic Fully-Connected Layers for Neural Networks [J].
Babiloni, Francesca ;
Tanay, Thomas ;
Deng, Jiankang ;
Maggioni, Matteo ;
Zafeiriou, Stefanos .
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, :1366-1375
[4]   Error-backpropagation in temporally encoded networks of spiking neurons [J].
Bohte, SM ;
Kok, JN ;
La Poutré, H .
NEUROCOMPUTING, 2002, 48 :17-37
[5]  
Cai WQ, 2023, Arxiv, DOI arXiv:2209.10837
[6]   Tensor Low-Rank Reconstruction for Semantic Segmentation [J].
Chen, Wanli ;
Zhu, Xinge ;
Sun, Ruoqi ;
He, Junjun ;
Li, Ruiyu ;
Shen, Xiaoyong ;
Yu, Bei .
COMPUTER VISION - ECCV 2020, PT XVII, 2020, 12362 :52-69
[7]  
Cheng X, 2020, PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P1519
[8]  
Deng S., 2021, ficient training of spiking neural network via gradient re
[9]  
Diehl PU, 2015, IEEE IJCNN
[10]  
Fang W., 2020, "Spikingjelly