ACE-SNN: Algorithm-Hardware Co-design of Energy-Efficient & Low-Latency Deep Spiking Neural Networks for 3D Image Recognition

被引:11
|
作者
Datta, Gourav [1 ]
Kundu, Souvik [1 ]
Jaiswal, Akhilesh R. [1 ]
Beerel, Peter A. [1 ]
机构
[1] Univ Southern Calif, Ming Hsieh Dept Elect & Comp Engn, Los Angeles, CA 90007 USA
关键词
hyperspectral images; spiking neural networks; quantization-aware; gradient descent; processing-in-memory; SRAM; CLASSIFICATION; ARCHITECTURE; ACCELERATOR; POWER;
D O I
10.3389/fnins.2022.815258
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
High-quality 3D image recognition is an important component of many vision and robotics systems. However, the accurate processing of these images requires the use of compute-expensive 3D Convolutional Neural Networks (CNNs). To address this challenge, we propose the use of Spiking Neural Networks (SNNs) that are generated from iso-architecture CNNs and trained with quantization-aware gradient descent to optimize their weights, membrane leak, and firing thresholds. During both training and inference, the analog pixel values of a 3D image are directly applied to the input layer of the SNN without the need to convert to a spike-train. This significantly reduces the training and inference latency and results in high degree of activation sparsity, which yields significant improvements in computational efficiency. However, this introduces energy-hungry digital multiplications in the first layer of our models, which we propose to mitigate using a processing-in-memory (PIM) architecture. To evaluate our proposal, we propose a 3D and a 3D/2D hybrid SNN-compatible convolutional architecture and choose hyperspectral imaging (HSI) as an application for 3D image recognition. We achieve overall test accuracy of 98.68, 99.50, and 97.95% with 5 time steps (inference latency) and 6-bit weight quantization on the Indian Pines, Pavia University, and Salinas Scene datasets, respectively. In particular, our models implemented using standard digital hardware achieved accuracies similar to state-of-the-art (SOTA) with ~560.6x and ~44.8x less average energy than an iso-architecture full-precision and 6-bit quantized CNN, respectively. Adopting the PIM architecture in the first layer, further improves the average energy, delay, and energy-delay-product (EDP) by 30, 7, and 38%, respectively.
引用
收藏
页数:21
相关论文
共 11 条
  • [1] STELLAR: Energy-Efficient and Low-Latency SNN Algorithm and Hardware Co-design with Spatiotemporal Computation
    Mao, Ruixin
    Tang, Lin
    Yuan, Xingyu
    Liu, Ye
    Zhou, Jun
    2024 IEEE INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE COMPUTER ARCHITECTURE, HPCA 2024, 2024, : 172 - 185
  • [2] Algorithm-hardware co-design for Energy-Efficient A/D conversion in ReRAM-based accelerators
    Zhang, Chenguang
    Yuan, Zhihang
    Li, Xingchen
    Sun, Guangyu
    2024 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2024,
  • [3] Scalable Energy-Efficient, Low-Latency Implementations of Trained Spiking Deep Belief Networks on SpiNNaker
    Stromatias, Evangelos
    Neil, Daniel
    Galluppi, Francesco
    Pfeiffer, Michael
    Liu, Shih-Chii
    Furber, Steve
    2015 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2015,
  • [4] A Low-Latency Framework With Algorithm-Hardware Co-Optimization for 3-D Point Cloud
    Yu, Yue
    Mao, Wendong
    Luo, Jiapeng
    Wang, Zhongfeng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2023, 70 (11) : 4221 - 4225
  • [5] Algorithm and Hardware Co-Design of Energy-Efficient LSTM Networks for Video Recognition With Hierarchical Tucker Tensor Decomposition
    Gong, Yu
    Yin, Miao
    Huang, Lingyi
    Deng, Chunhua
    Yuan, Bo
    IEEE TRANSACTIONS ON COMPUTERS, 2022, 71 (12) : 3101 - 3114
  • [6] Reinforcement co-Learning of Deep and Spiking Neural Networks for Energy-Efficient Mapless Navigation with Neuromorphic Hardware
    Tang, Guangzhi
    Kumar, Neelesh
    Michmizos, Konstantinos P.
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 6090 - 6097
  • [7] IMG-SMP: Algorithm and Hardware Co-Design for Real-time Energy-efficient Neural Motion Planning
    Huang, Lingyi
    Zang, Xiao
    Gong, Yu
    Deng, Chunhua
    Yi, Jingang
    Yuan, Bo
    PROCEEDINGS OF THE 32ND GREAT LAKES SYMPOSIUM ON VLSI 2022, GLSVLSI 2022, 2022, : 373 - 377
  • [8] Fixed-Point Quantization of 3D Convolutional Neural Networks for Energy-Efficient Action Recognition
    Lee, Hyunhoon
    Byun, Younghoon
    Hwang, Seokha
    Lee, Sunggu
    Lee, Youngjoo
    2018 INTERNATIONAL SOC DESIGN CONFERENCE (ISOCC), 2018, : 129 - 130
  • [9] An Energy-Efficient 3D Cross-Ring Accelerator With 3D-SRAM Cubes for Hybrid Deep Neural Networks
    Lu, Wei
    Huang, Po-Tsang
    Chen, Hung-Ming
    Hwang, Wei
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2021, 11 (04) : 776 - 788
  • [10] Hyperdimensional Computing with 3D VRRAM In-Memory Kernels: Device-Architecture Co-Design for Energy-Efficient, Error-Resilient Language Recognition
    Li, Haitong
    Wu, Tony F.
    Rahimi, Abbas
    Li, Kai-Shin
    Rusch, Miles
    Lin, Chang-Hsien
    Hsu, Juo-Luen
    Sabry, Mohamed M.
    Eryilmaz, S. Burc
    Sohn, Joon
    Chiu, Wen-Cheng
    Chen, Min-Cheng
    Wu, Tsung-Ta
    Shieh, Jia-Min
    Yeh, Wen-Kuan
    Rabaey, Jan M.
    Mitra, Subhasish
    Wong, H. -S. Philip
    2016 IEEE INTERNATIONAL ELECTRON DEVICES MEETING (IEDM), 2016,