High-performance Sparsity-aware NPU with Reconfigurable Comparator-multiplier Architecture

被引:0
作者
Ryu, Sungju [1 ]
Kim, Jae-Joon [1 ]
机构
[1] Sogang Univ, Dept Syst Semicond Engn, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
Neural processing unit; sparse matrix; multiplier; weight pruning; hardware accelerator;
D O I
10.5573/JSTS.2024.24.6.572
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
-Sparsity-aware neural processing units have been studied to exploit computational skipping on less important features in the neural network models. However, neural network layers typically show various matrix densities, so the hardware performance varies depending on the layer characteristics. In this paper, we introduce a reconfigurable comparator-multiplier architecture, so we can dynamically change the number of comparator/multiplier modules. The proposed reconfigurable architecture increases the throughput by 1.06-17.00x compared to the previous sparsity- aware hardware accelerators.
引用
收藏
页码:572 / 577
页数:6
相关论文
共 4 条
  • [1] Parashar A., Et al., Scnn: An accelerator for compressed-sparse convolutional neural networks, 2017 ACM/IEEE 44th Annual International Symposium on Computer Architecture (ISCA), pp. 27-40, (2017)
  • [2] Zhang J.-F., Et al., Snap: An efficient sparse neural acceleration processor for unstructured sparse deep neural network inference, IEEE Journal of SolidState Circuits, 56, 2, pp. 636-647, (2020)
  • [3] Gondimalla A., Et al., Sparten: A sparse tensor accelerator for convolutional neural networks, Proceedings of the 52nd Annual IEEE/ACM International Symposium on Microarchitecture, pp. 151-165, (2019)
  • [4] Ryu S., Et al., Sprite: sparsity-aware neural processing unit with constant probability of indexmatching, 2021 Design, Automation & Test in Europe Conference & Exhibition (DATE), pp. 663-666, (2021)