Low-Complexity Approximate Convolutional Neural Networks

被引:21
作者
Cintra, Renato J. [1 ,2 ,3 ]
Duffner, Stefan [4 ]
Garcia, Christophe [4 ]
Leite, Andre [2 ]
机构
[1] Univ Lyon, CNRS, INSA Lyon, LIRIS,UMR5205, F-69621 Lyon, France
[2] Univ Fed Pernambuco, Dept Estat, Signal Proc Grp, BR-50670901 Recife, PE, Brazil
[3] Univ Calgary, Dept Elect & Comp Engn, Calgary, AB T2N 1N4, Canada
[4] Univ Lyon, CNRS, INSA Lyon, LIRIS,UMR5205, F-69621 Villeurbanne, France
关键词
Approximation; convolutional neural networks (ConvNets); numerical computation; optimization; SIGMOID FUNCTION; INTEGER DCT; IMAGE; TRANSFORM; IMPLEMENTATION; ARCHITECTURE; DESIGN; TIME;
D O I
10.1109/TNNLS.2018.2815435
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we present an approach for minimizing the computational complexity of the trained convolutional neural networks (ConvNets). The idea is to approximate all elements of a given ConvNet and replace the original convolutional filters and parameters (pooling and bias coefficients; and activation function) with an efficient approximations capable of extreme reductions in computational complexity. Low-complexity convolution filters are obtained through a binary (zero and one) linear programming scheme based on the Frobenius norm over sets of dyadic rationals. The resulting matrices allow for multiplication-free computations requiring only addition and bit-shifting operations. Such low-complexity structures pave the way for low power, efficient hardware designs. We applied our approach on three use cases of different complexities: 1) a "light" but efficient ConvNet for face detection (with around 1000 parameters); 2) another one for hand-written digit classification (with more than 180000 parameters); and 3) a significantly larger ConvNet: AlexNet with approximate to 1.2 million matrices. We evaluated the overall performance on the respective tasks for different levels of approximations. In all considered applications, very low-complexity approximations have been derived maintaining an almost equal classification performance.
引用
收藏
页码:5981 / 5992
页数:12
相关论文
共 90 条
[1]   DISCRETE COSINE TRANSFORM [J].
AHMED, N ;
NATARAJAN, T ;
RAO, KR .
IEEE TRANSACTIONS ON COMPUTERS, 1974, C 23 (01) :90-93
[2]  
Alippi C., 1991, 1991 IEEE International Symposium on Circuits and Systems (Cat. No.91CH3006-4), P1505, DOI 10.1109/ISCAS.1991.176661
[3]   Piecewise linear approximation applied to nonlinear function of a neural network [J].
Amin, H ;
Curtis, KM ;
Hayes-Gill, BR .
IEE PROCEEDINGS-CIRCUITS DEVICES AND SYSTEMS, 1997, 144 (06) :313-317
[4]  
[Anonymous], P INT C CONN PERSP O
[5]  
[Anonymous], P INT C COMP VIS
[6]  
[Anonymous], MNIST DATA BASE HAND
[7]  
[Anonymous], 2004, LECT NOTES COMPUTER
[8]  
[Anonymous], 2019, R: A language for environment for statistical computing
[9]  
[Anonymous], 2013, INTERSPEECH
[10]  
[Anonymous], 2014, ARXIV13125851V5 CSCV