SaARSP: An Architecture for Systolic-Array Acceleration of Recurrent Spiking Neural Networks

被引:3
作者
Lee, Jeong-Jun [1 ]
Zhang, Wenrui [1 ]
Xie, Yuan [1 ]
Li, Peng [1 ]
机构
[1] Univ Calif Santa Barbara, Elect & Comp Engn, Santa Barbara, CA 93106 USA
基金
美国国家科学基金会;
关键词
Spiking neural networks; accelerators; computer architecture; IMPLEMENTATION; NEURONS; FLOW;
D O I
10.1145/3510854
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Spiking neural networks (SNNs) are brain-inspired event-driven models of computation with promising ultra-low energy dissipation. Rich network dynamics emergent in recurrent spiking neural networks (R-SNNs) can form temporally based memory, offering great potential in processing complex spatiotemporal data. However, recurrence in network connectivity produces tightly coupled data dependency in both space and time, rendering hardware acceleration of R-SNNs challenging. We present the first work to exploit spatiotemporal parallelisms to accelerate the R-SNN-based inference on systolic arrays using an architecture called SaARSP. We decouple the processing of feedforward synaptic connections from that of recurrent connections to allow for the exploitation of parallelisms across multiple time points. We propose a novel time window size optimization (TWSO) technique, to further explore the temporal granularity of the proposed decoupling in terms of optimal time window size and reconfiguration of the systolic array considering layer-dependent connectivity to boost performance. Stationary dataflow and time window size are jointly optimized to trade off between weight data reuse and movements of partial sums, the two bottlenecks in latency and energy dissipation of the accelerator. The proposed systolic-array architecture offers a unifying solution to an acceleration of both feedforward and recurrent SNNs, and delivers 4,000X EDP improvement on average for different R-SNN benchmarks over a conventional baseline.
引用
收藏
页数:23
相关论文
共 59 条
[1]   A digital hardware implementation of spiking neural networks with binary FORCE training [J].
Akbarzadeh-Sherbaf, Kaveh ;
Safari, Saeed ;
Vahabie, Abdol-Hossein .
NEUROCOMPUTING, 2020, 412 :129-142
[2]   True North: Design and Tool Flow of a 65 mW 1 Million Neuron Programmable Neurosynaptic Chip [J].
Akopyan, Filipp ;
Sawada, Jun ;
Cassidy, Andrew ;
Alvarez-Icaza, Rodrigo ;
Arthur, John ;
Merolla, Paul ;
Imam, Nabil ;
Nakamura, Yutaka ;
Datta, Pallab ;
Nam, Gi-Joon ;
Taba, Brian ;
Beakes, Michael ;
Brezzo, Bernard ;
Kuang, Jente B. ;
Manohar, Rajit ;
Risk, William P. ;
Jackson, Bryan ;
Modha, Dharmendra S. .
IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2015, 34 (10) :1537-1557
[3]  
[Anonymous], 1993, Tech. Rep. LDC93S1
[4]  
[Anonymous], 2018, Advances in Neural Information Processing Systems
[5]   Feature Representations for Neuromorphic Audio Spike Streams [J].
Anumula, Jithendar ;
Neil, Daniel ;
Delbruck, Tobi ;
Liu, Shih-Chii .
FRONTIERS IN NEUROSCIENCE, 2018, 12
[6]  
Bellec G, 2018, 32 C NEURAL INFORM P
[7]   Synaptic modifications in cultured hippocampal neurons: Dependence on spike timing, synaptic strength, and postsynaptic cell type [J].
Bi, GQ ;
Poo, MM .
JOURNAL OF NEUROSCIENCE, 1998, 18 (24) :10464-10472
[8]   Spiking Neural Networks Hardware Implementations and Challenges: A Survey [J].
Bouvier, Maxence ;
Valentian, Alexandre ;
Mesquida, Thomas ;
Rummens, Francois ;
Reyboz, Marina ;
Vianello, Elisa ;
Beigne, Edith .
ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2019, 15 (02)
[9]   Spiking Deep Convolutional Neural Networks for Energy-Efficient Object Recognition [J].
Cao, Yongqiang ;
Chen, Yang ;
Khosla, Deepak .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2015, 113 (01) :54-66
[10]   Spike timing-dependent plasticity: A Hebbian learning rule [J].
Caporale, Natalia ;
Dan, Yang .
ANNUAL REVIEW OF NEUROSCIENCE, 2008, 31 :25-46