A 71.2-μW Speech Recognition Accelerator With Recurrent Spiking Neural Network

被引:2
|
作者
Yang, Chih-Chyau [1 ,2 ]
Chang, Tian-Sheuan [1 ]
机构
[1] Natl Yang Ming Chiao Tung Univ, Inst Elect, Hsinchu 30010, Taiwan
[2] Taiwan Semicond Res Inst, Hsinchu 30078, Taiwan
关键词
Deep-learning accelerator; recurrent spiking neural networks; zero skipping hardware; ultra low power; model compression;
D O I
10.1109/TCSI.2024.3387993
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This paper introduces a 71.2- mu W speech recognition accelerator designed for edge devices' real-time applications, emphasizing an ultra low power design. Achieved through algorithm and hardware co-optimizations, we propose a compact recurrent spiking neural network with two recurrent layers, one fully connected layer, and a low time step (1 or 2). The 2.79-MB model undergoes pruning and 4-bit fixed-point quantization, shrinking it by 96.42% to 0.1 MB. On the hardware front, we take advantage of mixed-level pruning, zero-skipping and merged spike techniques, reducing complexity by 90.49% to 13.86 MMAC/S. The parallel time-step execution addresses inter-time-step data dependencies and enables weight buffer power savings through weight sharing. Capitalizing on the sparse spike activity, an input broadcasting scheme eliminates zero computations, further saving power. Implemented on the TSMC 28-nm process, the design operates in real time at 100 kHz, consuming 71.2 mu W, surpassing state-of-the-art designs. At 500 MHz, it has 28.41 TOPS/W and 1903.11 GOPS/mm(2) in energy and area efficiency, respectively.
引用
收藏
页码:3203 / 3213
页数:11
相关论文
共 50 条
  • [1] The spiking neural network based on fMRI for speech recognition
    Song, Yihua
    Guo, Lei
    Man, Menghua
    Wu, Youxi
    PATTERN RECOGNITION, 2024, 155
  • [2] Temporal Spiking Recurrent Neural Network for Action Recognition
    Wang, Wei
    Hao, Siyuan
    Wei, Yunchao
    Xia, Shengtao
    Feng, Jiashi
    Sebe, Nicu
    IEEE ACCESS, 2019, 7 : 117165 - 117175
  • [3] Speech emotion recognition based on spiking neural network and convolutional neural network
    Du, Chengyan
    Liu, Fu
    Kang, Bing
    Hou, Tao
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 147
  • [4] Stochastic Recurrent Neural Network for Speech Recognition
    Chien, Jen-Tzung
    Shen, Chen
    18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 1313 - 1317
  • [5] EEG Recognition of Epilepsy Based on Spiking Recurrent Neural Network
    Zhou, Shitao
    Liu, Yijun
    Ye, Wujian
    PROCEEDINGS OF 2024 4TH INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND INTELLIGENT COMPUTING, BIC 2024, 2024, : 127 - 132
  • [6] Dynamic Spatiotemporal Pattern Recognition With Recurrent Spiking Neural Network
    Shen, Jiangrong
    Liu, Jian K.
    Wang, Yueming
    NEURAL COMPUTATION, 2021, 33 (11) : 2971 - 2995
  • [7] A Scratchpad Spiking Neural Network Accelerator
    Karakchi, Rasha
    2024 IEEE 3RD INTERNATIONAL CONFERENCE ON COMPUTING AND MACHINE INTELLIGENCE, ICMI 2024, 2024,
  • [8] A spiking recurrent neural network
    Li, Y
    Harris, JG
    VLSI 2004: IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI, PROCEEDINGS, 2004, : 321 - 322
  • [9] Real-time Speech Recognition for IoT Purpose using a Delta Recurrent Neural Network Accelerator
    Gao, Chang
    Braun, Stefan
    Kiselev, Ilya
    Anumula, Jithendar
    Delbruck, Tobi
    Liu, Shih-Chii
    2019 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2019,
  • [10] An 8.93 TOPS/W LSTM Recurrent Neural Network Accelerator Featuring Hierarchical Coarse-Grain Sparsity for On-Device Speech Recognition
    Kadetotad, Deepak
    Yin, Shihui
    Berisha, Visar
    Chakrabarti, Chaitali
    Seo, Jae-sun
    IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2020, 55 (07) : 1877 - 1887