A comprehensive review of Binary Neural Network

被引:0
作者
Chunyu Yuan
Sos S. Agaian
机构
[1] City University of New York,The Graduate Center
[2] City University of New York,College of Staten Island
来源
Artificial Intelligence Review | 2023年 / 56卷
关键词
Binary Neural Network; Convolution Neural Network; Model compression and acceleration; Binarization; Quantization;
D O I
暂无
中图分类号
学科分类号
摘要
Deep learning (DL) has recently changed the development of intelligent systems and is widely adopted in many real-life applications. Despite their various benefits and potentials, there is a high demand for DL processing in different computationally limited and energy-constrained devices. It is natural to study game-changing technologies such as Binary Neural Networks (BNN) to increase DL capabilities. Recently remarkable progress has been made in BNN since they can be implemented and embedded on tiny restricted devices and save a significant amount of storage, computation cost, and energy consumption. However, nearly all BNN acts trade with extra memory, computation cost, and higher performance. This article provides a complete overview of recent developments in BNN. This article focuses exclusively on 1-bit activations and weights 1-bit convolution networks, contrary to previous surveys in which low-bit works are mixed in. It conducted a complete investigation of BNN’s development—from their predecessors to the latest BNN algorithms/techniques, presenting a broad design pipeline and discussing each module’s variants. Along the way, it examines BNN (a) purpose: their early successes and challenges; (b) BNN optimization: selected representative works that contain essential optimization techniques; (c) deployment: open-source frameworks for BNN modeling and development; (d) terminal: efficient computing architectures and devices for BNN and (e) applications: diverse applications with BNN. Moreover, this paper discusses potential directions and future research opportunities in each section.
引用
收藏
页码:12949 / 13013
页数:64
相关论文
共 184 条
  • [1] Blott M(2018)FINN-R: an end-to-end deep-learning framework for fast exploration of quantized neural networks ACM Trans Reconfig Technol Syst 11 1-23
  • [2] Preußer TB(2019)Salient object detection: a survey Comput Vis Media 5 117-150
  • [3] Fraser NJ(1996)Bagging predictors Mach Learn 24 123-140
  • [4] Gambardella G(2001)A survey of free-form object representation and recognition techniques Comput Vis Image Underst 81 166-210
  • [5] O’Brien K(2011)A survey on artificial intelligence approaches for medical image classification Indian J Sci Technol 4 1583-1595
  • [6] Umuroglu Y(2019)Stochastic quantization for learning accurate low-bit deep neural networks Int J Comput Vis 127 1629-1642
  • [7] Leeser M(2010)The pascal visual object classes (VOC) challenge Int J Comput Vis 88 303-338
  • [8] Vissers K(1997)A decision-theoretic generalization of on-line learning and an application to boosting J Comput Syst Sci 55 119-139
  • [9] Borji A(2020)Riptide: fast end-to-end binarized neural networks Proc Mach Learn Syst 2 379-389
  • [10] Cheng MM(2020)O3BNN-R: an out-of-order architecture for high-performance and regularized BNN inference IEEE Trans Parallel Distrib Syst 32 199-213