Learning the Sparsity for ReRAM: Mapping and Pruning Sparse Neural Network for ReRAM based Accelerator

被引:63
作者
Lin, Jilan [1 ,2 ]
Zhu, Zhenhua [2 ]
Wang, Yu [2 ]
Xie, Yuan [1 ]
机构
[1] UCSB, Dept Elect & Comp Engn, Santa Barbara, CA 93106 USA
[2] Tsinghua Univ, Dept Elect Engn, Beijing, Peoples R China
来源
24TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC 2019) | 2019年
关键词
D O I
10.1145/3287624.3287715
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
With the in-memory processing ability, ReRAM based computing gets more and more attractive for accelerating neural networks (NNs). However, most ReRAM based accelerators cannot support efficient mapping for sparse NN, and we need to map the whole dense matrix onto ReRAM crossbar array to achieve O(1) computation complexity. In this paper, we propose a sparse NN mapping scheme based on elements clustering to achieve better ReRAM crossbar utilization. Further, we propose crossbar-grained pruning algorithm to remove the crossbars with low utilization. Finally, since most current ReRAM devices cannot achieve high precision, we analyze the effect of quantization precision for sparse NN, and propose to complete high-precision composing in the analog field and design related periphery circuits. In our experiments, we discuss how the system performs with different crossbar sizes to choose the optimized design. Our results show that our mapping scheme for sparse NN with proposed pruning algorithm achieves 3 - 5x energy efficiency and more than 2:5 - 6x speedup, compared with those accelerators for dense NN. Also, the accuracy experiments show that our pruning method appears to have almost no accuracy loss.
引用
收藏
页码:639 / 644
页数:6
相关论文
共 21 条
[1]  
[Anonymous], ICCAD
[2]  
[Anonymous], 2018, DAC
[3]  
[Anonymous], VLSI
[4]  
[Anonymous], 2016, LEARNING STRUCTURED
[5]  
[Anonymous], 2017, IEEE CVPRW, DOI DOI 10.1186/S12888-017-1313-3
[6]  
[Anonymous], DAC
[7]  
[Anonymous], 2016, DAC 16
[8]  
Chen WH, 2018, ISSCC DIG TECH PAP I, P494, DOI 10.1109/ISSCC.2018.8310400
[9]   PRIME: A Novel Processing-in-memory Architecture for Neural Network Computation in ReRAM-based Main Memory [J].
Chi, Ping ;
Li, Shuangchen ;
Xu, Cong ;
Zhang, Tao ;
Zhao, Jishen ;
Liu, Yongpan ;
Wang, Yu ;
Xie, Yuan .
2016 ACM/IEEE 43RD ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA), 2016, :27-39
[10]   NVSim: A Circuit-Level Performance, Energy, and Area Model for Emerging Nonvolatile Memory [J].
Dong, Xiangyu ;
Xu, Cong ;
Xie, Yuan ;
Jouppi, Norman P. .
IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2012, 31 (07) :994-1007