CyNAPSE: A Low-power Reconfigurable Neural Inference Accelerator for Spiking Neural Networks

被引:2
作者
Saha, Saunak [1 ]
Duwe, Henry [1 ]
Zambreno, Joseph [1 ]
机构
[1] Iowa State Univ, Dept Elect & Comp Engn, Ames, IA 50011 USA
来源
JOURNAL OF SIGNAL PROCESSING SYSTEMS FOR SIGNAL IMAGE AND VIDEO TECHNOLOGY | 2020年 / 92卷 / 09期
基金
美国国家科学基金会;
关键词
Neuromorphic; Spiking neural networks; Reconfigurable; Accelerator; Memory; Caching; Leakage; Energy efficiency; PROCESSOR; MODEL; ARCHITECTURE;
D O I
10.1007/s11265-020-01546-x
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
While neural network models keep scaling in depth and computational requirements, biologically accurate models are becoming more interesting for low-cost inference. Coupled with the need to bring more computation to the edge in resource-constrained embedded and IoT devices, specialized ultra-low power accelerators for spiking neural networks are being developed. Having a large variance in the models employed in these networks, these accelerators need to be flexible, user-configurable, performant and energy efficient. In this paper, we describe CyNAPSE, a fully digital accelerator designed to emulate neural dynamics of diverse spiking networks. Since the use case of our implementation is primarily concerned with energy efficiency, we take a closer look at the factors that could improve its energy consumption. We observe that while majority of its dynamic power consumption can be credited to memory traffic, its on-chip components suffer greatly from static leakage. Given that the event-driven spike processing algorithm is naturally memory-intensive and has a large number of idle processing elements, it makes sense to tackle each of these problems towards a more efficient hardware implementation. With a diverse set of network benchmarks, we incorporate a detailed study of memory patterns that ultimately informs our choice of an application-specific network-adaptive memory management strategy to reduce dynamic power consumption of the chip. Subsequently, we also propose and evaluate a leakage mitigation strategy for runtime control of idle power. Using both the RTL implementation and a software simulation of CyNAPSE, we measure the relative benefits of these undertakings. Results show that our adaptive memory management policy results in up to 22% more reduction in dynamic power consumption compared to conventional policies. The runtime leakage mitigation techniques show that up to 99.92% and at least 14% savings in leakage energy consumption is achievable in CyNAPSE hardware modules.
引用
收藏
页码:907 / 929
页数:23
相关论文
共 64 条
  • [31] Which model to use for cortical spiking neurons?
    Izhikevich, EM
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 2004, 15 (05): : 1063 - 1070
  • [32] Jiang HL, 2005, PR IEEE COMP DESIGN, P559
  • [33] Generalized integrate-and-fire models of neuronal activity approximate spike trains of a detailed model to a high degree of accuracy
    Jolivet, R
    Lewis, TJ
    Gerstner, W
    [J]. JOURNAL OF NEUROPHYSIOLOGY, 2004, 92 (02) : 959 - 976
  • [34] Jolivet R., 2006, Advances in Neural Information Processing Systems, P595
  • [35] In-Datacenter Performance Analysis of a Tensor Processing Unit
    Jouppi, Norman P.
    Young, Cliff
    Patil, Nishant
    Patterson, David
    Agrawal, Gaurav
    Bajwa, Raminder
    Bates, Sarah
    Bhatia, Suresh
    Boden, Nan
    Borchers, Al
    Boyle, Rick
    Cantin, Pierre-luc
    Chao, Clifford
    Clark, Chris
    Coriell, Jeremy
    Daley, Mike
    Dau, Matt
    Dean, Jeffrey
    Gelb, Ben
    Ghaemmaghami, Tara Vazir
    Gottipati, Rajendra
    Gulland, William
    Hagmann, Robert
    Ho, C. Richard
    Hogberg, Doug
    Hu, John
    Hundt, Robert
    Hurt, Dan
    Ibarz, Julian
    Jaffey, Aaron
    Jaworski, Alek
    Kaplan, Alexander
    Khaitan, Harshit
    Killebrew, Daniel
    Koch, Andy
    Kumar, Naveen
    Lacy, Steve
    Laudon, James
    Law, James
    Le, Diemthu
    Leary, Chris
    Liu, Zhuyuan
    Lucke, Kyle
    Lundin, Alan
    MacKean, Gordon
    Maggiore, Adriana
    Mahony, Maire
    Miller, Kieran
    Nagarajan, Rahul
    Narayanaswami, Ravi
    [J]. 44TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA 2017), 2017, : 1 - 12
  • [36] Kaxiras S, 2001, ACM COMP AR, P240, DOI 10.1109/ISCA.2001.937453
  • [37] Khan S. M., 2010, Proceedings 2010 43rd Annual IEEE/ACM International Symposium on Microarchitecture (MICRO 2010), P175, DOI 10.1109/MICRO.2010.24
  • [38] A Reconfigurable Digital Neuromorphic Processor with Memristive Synaptic Crossbar for Cognitive Computing
    Kim, Yongtae
    Zhang, Yong
    Li, Peng
    [J]. ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2015, 11 (04)
  • [39] Ramulator: A Fast and Extensible DRAM Simulator
    Kim, Yoongu
    Yang, Weikun
    Mutlu, Onur
    [J]. IEEE COMPUTER ARCHITECTURE LETTERS, 2016, 15 (01) : 45 - 49
  • [40] Koch C., 1998, METHODS NEURONAL MOD