HEIF: Highly Efficient Stochastic Computing-Based Inference Framework for Deep Neural Networks

被引:54
|
作者
Li, Zhe [1 ]
Li, Ji [2 ]
Ren, Ao [1 ]
Cai, Ruizhe [1 ]
Ding, Caiwen [1 ]
Qian, Xuehai [2 ]
Draper, Jeffrey [2 ]
Yuan, Bo [3 ]
Tang, Jian [1 ]
Qiu, Qinru [1 ]
Wang, Yanzhi [4 ]
机构
[1] Syracuse Univ, Dept Elect Engn & Comp Sci, Syracuse, NY 13244 USA
[2] Univ Southern Calif, Dept Elect Engn, Los Angeles, CA 90089 USA
[3] CUNY City Coll, Dept Elect Engn, New York, NY 10031 USA
[4] Northeastern Univ, Dept Elect & Comp Engn, Boston, MA 02115 USA
关键词
ASIC; convolutional neural network; deep learning; energy-efficient; optimization; stochastic computing (SC);
D O I
10.1109/TCAD.2018.2852752
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep convolutional neural networks (DCNNs) are one of the most promising deep learning techniques and have been recognized as the dominant approach for almost all recognition and detection tasks. The computation of DCNNs is memory intensive due to large feature maps and neuron connections, and the performance highly depends on the capability of hardware resources. With the recent trend of wearable devices and Internet of Things, it becomes desirable to integrate the DCNNs onto embedded and portable devices that require low power and energy consumptions and small hardware footprints. Recently stochastic computing (SC)-DCNN demonstrated that SC as a low-cost substitute to binary-based computing radically simplifies the hardware implementation of arithmetic units and has the potential to satisfy the stringent power requirements in embedded devices. In SC, many arithmetic operations that are resource-consuming in binary designs can be implemented with very simple hardware logic, alleviating the extensive computational complexity. It offers a colossal design space for integration and optimization due to its reduced area and soft error resiliency. In this paper, we present HEIF, a highly efficient SC-based inference framework of the large-scale DCNNs, with broad applications including (but not limited to) LeNet-5 and AlexNet, that achieves high energy efficiency and low area/ hardware cost. Compared to SC-DCNN, HEIF features: 1) the first (to the best of our knowledge) SC-based rectified linear unit activation function to catch up with the recent advances in software models and mitigate degradation in application-level accuracy; 2) the redesigned approximate parallel counter and optimized stochastic multiplication using transmission gates and inverse mirror adders; and 3) the new optimization of weight storage using clustering. Most importantly, to achieve maximum energy efficiency while maintaining acceptable accuracy, HEIF considers holistic optimizations on cascade connection of function blocks in DCNN, pipelining technique, and bit-stream length reduction. Experimental results show that in large-scale applications HEIF outperforms previous SC-DCNN by the throughput of 4.1x, by area efficiency of up to 6.5x, and achieves up to 5.6x energy improvement.
引用
收藏
页码:1543 / 1556
页数:14
相关论文
共 50 条
  • [31] VLSI Implementation of Deep Neural Networks Using Integral Stochastic Computing
    Ardakani, Arash
    Leduc-Primeau, Fracois
    Onizawa, Naoya
    Hanyu, Takahiro
    Gross, Warren J.
    2016 9TH INTERNATIONAL SYMPOSIUM ON TURBO CODES AND ITERATIVE INFORMATION PROCESSING (ISTC), 2016, : 216 - 220
  • [32] Towards Acceleration of Deep Convolutional Neural Networks using Stochastic Computing
    Li, Ji
    Ren, Ao
    Li, Zhe
    Ding, Caiwen
    Yuan, Bo
    Qiu, Qinru
    Wang, Yanzhi
    2017 22ND ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2017, : 115 - 120
  • [33] DSCNN: Hardware-Oriented Optimization for Stochastic Computing Based Deep Convolutional Neural Networks
    Li, Zhe
    Ren, Ao
    Li, Ji
    Qiu, Qinru
    Wang, Yanzhi
    Yuan, Bo
    PROCEEDINGS OF THE 34TH IEEE INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD), 2016, : 678 - 681
  • [34] Hardware-Driven Nonlinear Activation for Stochastic Computing Based Deep Convolutional Neural Networks
    Li, Ji
    Yuan, Zihao
    Li, Zhe
    Ding, Caiwen
    Ren, Ao
    Qiu, Qinru
    Draper, Jeffrey
    Wang, Yanzhi
    2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 1230 - 1236
  • [35] Energy Efficient Stochastic-Based Deep Spiking Neural Networks for Sparse Datasets
    Alawad, Mohammed
    Yoon, Hong-Jun
    Tourassi, Georgia
    2017 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2017, : 311 - 318
  • [36] Efficient Hardware Acceleration for Approximate Inference of Bitwise Deep Neural Networks
    Vogel, Sebastian
    Guntoro, Andre
    Ascheid, Gerd
    2017 CONFERENCE ON DESIGN AND ARCHITECTURES FOR SIGNAL AND IMAGE PROCESSING (DASIP), 2017,
  • [37] Efficient Distributed Inference of Deep Neural Networks via Restructuring and Pruning
    Abdi, Afshin
    Rashidi, Saeed
    Fekri, Faramarz
    Krishna, Tushar
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 6, 2023, : 6640 - 6648
  • [38] Private Inference for Deep Neural Networks: A Secure, Adaptive, and Efficient Realization
    Cheng, Ke
    Xi, Ning
    Liu, Ximeng
    Zhu, Xinghui
    Gao, Haichang
    Zhang, Zhiwei
    Shen, Yulong
    IEEE TRANSACTIONS ON COMPUTERS, 2023, 72 (12) : 3519 - 3531
  • [39] Efficient Priors for Scalable Variational Inference in Bayesian Deep Neural Networks
    Krishnan, Ranganath
    Subedar, Mahesh
    Tickoo, Omesh
    Labs, Intel
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 773 - 777
  • [40] Inference and Energy Efficient Design of Deep Neural Networks for Embedded Devices
    Galanis, Ioannis
    Anagnostopoulos, Iraklis
    Nguyen, Chinh
    Bares, Guillermo
    Burkard, Dona
    2020 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI 2020), 2020, : 36 - 41