A Mixed Signal Architecture for Convolutional Neural Networks

被引:20
作者
Lou, Qiuwen [1 ]
Pan, Chenyun [2 ]
McGuinness, John [1 ]
Horvath, Andras [3 ]
Naeemi, Azad [4 ]
Niemier, Michael [1 ]
Hu, X. Sharon [1 ]
机构
[1] Univ Notre Dame, 100 Notre Dame Ave, Notre Dame, IN USA
[2] Univ Kensas, Lawrence, KS USA
[3] Pazmany Peter Catholic Univ, Szentkiralyi U 28, H-1008 Budapest, Hungary
[4] Georgia Inst Technol, Atlanta, GA 30332 USA
基金
美国国家科学基金会;
关键词
Hardware accelerator; convolutional neural networks; analog circuits; CNN; MACHINE; DESIGN; ENERGY;
D O I
10.1145/3304110
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural network (DNN) accelerators with improved energy and delay are desirable for meeting the requirements of hardware targeted for IoT and edge computing systems. Convolutional neural networks (CoNNs) belong to one of the most popular types of DNN architectures. This article presents the design and evaluation of an accelerator for CoNNs. The system-level architecture is based on mixed-signal, cellular neural networks (CeNNs). Specifically, we present (i) the implementation of different layers, including convolution, ReLU, and pooling, in a CoNN using CeNN, (ii) modified CoNN structures with CeNN-friendly layers to reduce computational overheads typically associated with a CoNN, (iii) a mixed-signal CeNN architecture that performs CoNN computations in the analog and mixed signal domain, and (iv) design space exploration that identifies what CeNN-based algorithm and architectural features fare best compared to existing algorithms and architectures when evaluated over common datasets-MNIST and CIFAR-10. Notably, the proposed approach can lead to 8.7x improvements in energy-delay product (EDP) per digit classification for the MNIST dataset at iso-accuracy when compared with the state-of-the-art DNN engine, while our approach could offer 4.3x improvements in EDP when compared to other network implementations for the CIFAR-10 dataset.
引用
收藏
页数:26
相关论文
共 63 条
[1]  
[Anonymous], P 3 INT C LEARNING R
[2]  
[Anonymous], 2016, SQUEEZENET ALEXNET L
[3]  
[Anonymous], 2010, MNIST HANDWRITTEN DI
[4]  
[Anonymous], 2009, MT-003 Tutorial
[5]  
[Anonymous], T ARCH CODE OPTIM
[6]  
[Anonymous], 2015, NIPS'15
[7]  
[Anonymous], 2015, Nature, DOI [10.1038/nature14539, DOI 10.1038/NATURE14539]
[8]  
[Anonymous], 2017, ICCAD-IEEE ACM INT
[9]  
[Anonymous], 2010, P ICML 10 P 27 INT C
[10]  
[Anonymous], 2017, ARXIV170404861