UCNN: Exploiting Computational Reuse in Deep Neural Networks via Weight Repetition

被引:129
作者
Hegde, Kartik [1 ]
Yu, Jiyong [1 ]
Agrawal, Rohit [1 ]
Yan, Mengjia [1 ]
Pellauer, Michael [2 ]
Fletcher, Christopher W. [1 ]
机构
[1] Univ Illinois, Urbana, IL 61801 USA
[2] NVIDIA, Santa Clara, CA USA
来源
2018 ACM/IEEE 45TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA) | 2018年
关键词
D O I
10.1109/ISCA.2018.00062
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional Neural Networks (CNNs) have begun to permeate all corners of electronic society (from voice recognition to scene generation) due to their high accuracy and machine efficiency per operation. At their core, CNN computations are made up of multi-dimensional dot products between weight and input vectors. This paper studies how weight repetition when the same weight occurs multiple times in or across weight vectors-can be exploited to save energy and improve performance during CNN inference. This generalizes a popular line of work to improve efficiency from CNN weight sparsity, as reducing computation due to repeated zero weights is a special case of reducing computation due to repeated weights. To exploit weight repetition, this paper proposes a new CNN accelerator called the Unique Weight CNN Accelerator (UCNN). UCNN uses weight repetition to reuse CNN sub-computations (e.g., dot products) and to reduce CNN model size when stored in off-chip DRAM both of which save energy. UCNN further improves performance by exploiting sparsity in weights. We evaluate UCNN with an accelerator-level cycle and energy model and with an RTL implementation of the UCNN processing element. On three contemporary CNNs, UCNN improves throughput normalized energy consumption by 1.2x 4x, relative to a similarly provisioned baseline accelerator that uses Eyeriss-style sparsity optimizations. At the same time, the UCNN processing element adds only 17-24% area overhead relative to the same baseline.
引用
收藏
页码:674 / 687
页数:14
相关论文
共 32 条
[1]  
Albericio J., ISCA 16
[2]  
[Anonymous], CVPR 15
[3]  
Chen Y., MICRO 14
[4]  
Chen Y.-H., ISCA 16
[5]  
Ciregan D., CVPR 12
[6]  
Cong Jason, 2014, Artificial Neural Networks and Machine Learning - ICANN 2014. 24th International Conference on Artificial Neural Networks. Proceedings: LNCS 8681, P281, DOI 10.1007/978-3-319-11179-7_36
[7]  
Courbariaux M., NIPS 16
[8]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[9]  
Han S., NIPS 15
[10]  
Han S., ICLR 16