Adaptive Integer Quantisation for Convolutional Neural Networks through Evolutionary Algorithms

被引:0
作者
Wang, Ziwei [1 ]
Trefzer, Martin A. [1 ]
Bale, Simon J. [1 ]
Tyrrell, Andy M. [1 ]
机构
[1] Univ York, Dept Elect Engn, York, N Yorkshire, England
来源
2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021) | 2021年
关键词
Convolutional Neural Networks; Quantisation; Evolutionary Algorithms; Deep Learning;
D O I
10.1109/SSCI50451.2021.9659887
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
State-of-the-art Convolutional Neural Networks (CNNs) have become increasingly accurate. However, hundreds or thousands of megabytes data are involved to store them, making these networks also computationally expensive. For certain applications, such as Internet-of-Things (IoT), where such CNNs are to be implemented on resource-constrained and memory-constrained platforms, including Field-Programmable Gate Arrays (FPGAs) and embedded devices, CNN architectures and parameters have to be small and efficient. In this paper, an evolutionary algorithm (EA) based adaptive integer quantisation method is proposed to reduce network size. The proposed method uses single objective rank-based evolutionary strategy to find the best quantisation bin boundary for fixed quantised bit width. The performance of the proposed method is evaluated on a small CNN, the LeNet-5 architecture, using the CIFAR-10 dataset. The aim is to devise a methodology that allows adaptive quantisation of both weights and bias from 32-bit floating point to 8-bit integer representation for LeNet-5, while retaining accuracy. The experiments compare straight-forward (linear) quantisation from 32-bits to 8-bits with the proposed adaptive quantisation method. The results show that the proposed method is capable of quantising CNNs to lower bit width representation with only a slight loss in classification accuracy.
引用
收藏
页数:7
相关论文
共 30 条
[21]  
Stanley K. O., 2004, THESIS
[22]  
Su J., 2018, TECHNICAL REPORT
[23]   A Genetic Programming Approach to Designing Convolutional Neural Network Architectures [J].
Suganuma, Masanori ;
Shirakawa, Shinichi ;
Nagao, Tomoharu .
PROCEEDINGS OF THE 2017 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE (GECCO'17), 2017, :497-504
[24]  
Szegedy C, 2015, PROC CVPR IEEE, P1, DOI 10.1109/CVPR.2015.7298594
[25]  
Trefzer M. A., 2010, P 12 ANN C GEN EV CO, P595, DOI DOI 10.1145/1830483.1830593
[26]   FINN: A Framework for Fast, Scalable Binarized Neural Network Inference [J].
Umuroglu, Yaman ;
Fraser, Nicholas J. ;
Gambardella, Giulio ;
Blott, Michaela ;
Leong, Philip ;
Jahre, Magnus ;
Vissers, Kees .
FPGA'17: PROCEEDINGS OF THE 2017 ACM/SIGDA INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE GATE ARRAYS, 2017, :65-74
[27]  
Wang ZW, 2019, 2019 14TH INTERNATIONAL SYMPOSIUM ON RECONFIGURABLE COMMUNICATION-CENTRIC SYSTEMS-ON-CHIP (RECOSOC 2019), P35, DOI [10.1109/ReCoSoC48741.2019.9034956, 10.1109/recosoc48741.2019.9034956]
[28]   Quantized Convolutional Neural Networks for Mobile Devices [J].
Wu, Jiaxiang ;
Leng, Cong ;
Wang, Yuhang ;
Hu, Qinghao ;
Cheng, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :4820-4828
[29]   Genetic CNN [J].
Xie, Lingxi ;
Yuille, Alan .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :1388-1397
[30]  
Zhou Shuchang, 2016, DoReFa-Net: training low bitwidth convolutional neural networks with low bitwidth gradients