A Low Power and Low Latency FPGA-Based Spiking Neural Network Accelerator

被引:8
|
作者
Liu, Hanwen [1 ]
Chen, Yi [1 ]
Zeng, Zihang [2 ]
Zhang, Malu [1 ]
Qu, Hong [1 ]
机构
[1] Univ Elect Sci & Technol China, Dept Comp Sci & Engn, Chengdu, Peoples R China
[2] Univ Elect Sci & Technol China, Glasgow Coll, Chengdu, Peoples R China
来源
2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN | 2023年
基金
美国国家科学基金会;
关键词
Spiking Neural Networks; FPGA; Neuromorphic Accelerator; ON-CHIP; IMPLEMENTATION; PROCESSOR; SYSTEM;
D O I
10.1109/IJCNN54540.2023.10191153
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Spiking Neural Networks (SNNs), known as the third generation of the neural network, are famous for their biological plausibility and brain-like characteristics. Recent efforts further demonstrate the potential of SNNs in high-speed inference by designing accelerators with the parallelism of temporal or spatial dimensions. However, with the limitation of hardware resources, the accelerator designs must utilize off-chip memory to store many intermediate data, which leads to both high power consumption and long latency. In this paper, we focus on the data flow between layers to improve arithmetic efficiency. Based on the spike discrete property, we design a convolution-pooling(CONVP) unit that fuses the processing of the convolutional layer and pooling layer to reduce latency and resource utilization. Furthermore, for the fully-connected layer, we apply intra-output parallelism and inter-output parallelism to accelerate network inference. We demonstrate the effectiveness of our proposed hardware architecture by implementing different SNN models with the different datasets on a Zynq XA7Z020 FPGA. The experiments show that our accelerator can achieve about x28 inference speed up with a competitive power compared with FPGA implementation on MNIST dataset and a x15 inference speed up with low power compared with ASIC design on DVSGesture dataset.
引用
收藏
页数:8
相关论文
empty
未找到相关数据