ACE-SNN: Algorithm-Hardware Co-design of Energy-Efficient & Low-Latency Deep Spiking Neural Networks for 3D Image Recognition

被引:15
作者
Datta, Gourav [1 ]
Kundu, Souvik [1 ]
Jaiswal, Akhilesh R. [1 ]
Beerel, Peter A. [1 ]
机构
[1] Univ Southern Calif, Ming Hsieh Dept Elect & Comp Engn, Los Angeles, CA 90007 USA
关键词
hyperspectral images; spiking neural networks; quantization-aware; gradient descent; processing-in-memory; SRAM; CLASSIFICATION; ARCHITECTURE; ACCELERATOR; POWER;
D O I
10.3389/fnins.2022.815258
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
High-quality 3D image recognition is an important component of many vision and robotics systems. However, the accurate processing of these images requires the use of compute-expensive 3D Convolutional Neural Networks (CNNs). To address this challenge, we propose the use of Spiking Neural Networks (SNNs) that are generated from iso-architecture CNNs and trained with quantization-aware gradient descent to optimize their weights, membrane leak, and firing thresholds. During both training and inference, the analog pixel values of a 3D image are directly applied to the input layer of the SNN without the need to convert to a spike-train. This significantly reduces the training and inference latency and results in high degree of activation sparsity, which yields significant improvements in computational efficiency. However, this introduces energy-hungry digital multiplications in the first layer of our models, which we propose to mitigate using a processing-in-memory (PIM) architecture. To evaluate our proposal, we propose a 3D and a 3D/2D hybrid SNN-compatible convolutional architecture and choose hyperspectral imaging (HSI) as an application for 3D image recognition. We achieve overall test accuracy of 98.68, 99.50, and 97.95% with 5 time steps (inference latency) and 6-bit weight quantization on the Indian Pines, Pavia University, and Salinas Scene datasets, respectively. In particular, our models implemented using standard digital hardware achieved accuracies similar to state-of-the-art (SOTA) with ~560.6x and ~44.8x less average energy than an iso-architecture full-precision and 6-bit quantized CNN, respectively. Adopting the PIM architecture in the first layer, further improves the average energy, delay, and energy-delay-product (EDP) by 30, 7, and 38%, respectively.
引用
收藏
页数:21
相关论文
共 76 条
[31]   Reinforcement Learning to Rank in E-Commerce Search Engine: Formalization, Analysis, and Application [J].
Hu, Yujing ;
Da, Qing ;
Zeng, Anxiang ;
Yu, Yang ;
Xu, Yinghui .
KDD'18: PROCEEDINGS OF THE 24TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2018, :368-377
[32]  
Jain Sarthak, 2020, ARXIV PREPRINT ARXIV
[33]   8T SRAM Cell as a Multibit Dot-Product Engine for Beyond Von Neumann Computing [J].
Jaiswal, Akhilesh ;
Chakraborty, Indranil ;
Agrawal, Amogh ;
Roy, Kaushik .
IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2019, 27 (11) :2556-2567
[34]   An In-Memory VLSI Architecture for Convolutional Neural Networks [J].
Kang, Mingu ;
Lim, Sungmin ;
Gonugondla, Sujan ;
Shanbhag, Naresh R. .
IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2018, 8 (03) :494-505
[35]   Temporal Backpropagation for Spiking Neural Networks with One Spike per Neuron [J].
Kheradpisheh, Saeed Reza ;
Masquelier, Timothee .
INTERNATIONAL JOURNAL OF NEURAL SYSTEMS, 2020, 30 (06)
[36]   Deep neural networks with weighted spikes [J].
Kim, Jaehyun ;
Kim, Heesu ;
Huh, Subin ;
Lee, Jinho ;
Choi, Kiyoung .
NEUROCOMPUTING, 2018, 311 :373-386
[37]   Sparse multinomial logistic regression: Fast algorithms and generalization bounds [J].
Krishnapuram, B ;
Carin, L ;
Figueiredo, MAT ;
Hartemink, AJ .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2005, 27 (06) :957-968
[38]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[39]  
Kundu S., 2021, ARXIV PREPRINT ARXIV
[40]   HIRE-SNN: Harnessing the Inherent Robustness of Energy-Efficient Deep Spiking Neural Networks by Training with Crafted Input Noise [J].
Kundu, Souvik ;
Pedram, Massoud ;
Beerel, Peter A. .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :5189-5198