STCA-SNN: self-attention-based temporal-channel joint attention for spiking neural networks

被引:4
作者
Wu, Xiyan [1 ]
Song, Yong [1 ]
Zhou, Ya [1 ]
Jiang, Yurong [1 ]
Bai, Yashuo [1 ]
Li, Xinyi [1 ]
Yang, Xin [1 ]
机构
[1] Beijing Inst Technol, Sch Opt & Photon, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
spiking neural networks; self-attention; temporal-channel; neuromorphic computing; event streams; DEEPER;
D O I
10.3389/fnins.2023.1261543
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Spiking Neural Networks (SNNs) have shown great promise in processing spatio-temporal information compared to Artificial Neural Networks (ANNs). However, there remains a performance gap between SNNs and ANNs, which impedes the practical application of SNNs. With intrinsic event-triggered property and temporal dynamics, SNNs have the potential to effectively extract spatio-temporal features from event streams. To leverage the temporal potential of SNNs, we propose a self-attention-based temporal-channel joint attention SNN (STCA-SNN) with end-to-end training, which infers attention weights along both temporal and channel dimensions concurrently. It models global temporal and channel information correlations with self-attention, enabling the network to learn 'what' and 'when' to attend simultaneously. Our experimental results show that STCA-SNNs achieve better performance on N-MNIST (99.67%), CIFAR10-DVS (81.6%), and N-Caltech 101 (80.88%) compared with the state-of-the-art SNNs. Meanwhile, our ablation study demonstrates that STCA-SNNs improve the accuracy of event stream classification tasks.
引用
收藏
页数:10
相关论文
共 61 条
[1]  
Amir A., 2017, P IEEE C COMPUTER VI
[2]  
Ba J., 2014, Multiple object recognition with visual attention
[3]  
Bellec G, 2018, ADV NEUR IN, V31
[4]  
Bu T, 2022, OPTIMIZED POTENTIAL
[5]   A Spatial-Channel-Temporal-Fused Attention for Spiking Neural Networks [J].
Cai, Wuque ;
Sun, Hongze ;
Liu, Rui ;
Cui, Yan ;
Wang, Jun ;
Xia, Yang ;
Yao, Dezhong ;
Guo, Daqing .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (10) :14315-14329
[6]   Spiking Deep Convolutional Neural Networks for Energy-Efficient Object Recognition [J].
Cao, Yongqiang ;
Chen, Yang ;
Khosla, Deepak .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2015, 113 (01) :54-66
[7]   DET: A High-resolution DVS Dataset for Lane Extraction [J].
Cheng, Wensheng ;
Luo, Hao ;
Yang, Wen ;
Yu, Lei ;
Chen, Shoushun ;
Li, Wei .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, :1666-1675
[8]   Unsupervised learning of digit recognition using spike-timing-dependent plasticity [J].
Diehl, Peter U. ;
Cook, Matthew .
FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2015, 9
[9]  
Ding Jianhao, 2021, INT JOINT C ART INT
[10]  
Dosovitskiy Alexey., 2021, PROC INT C LEARN REP, P2021, DOI [10.48550/arXiv.2010.11929, DOI 10.48550/ARXIV.2010.11929]