A Resource-Efficient Convolutional Neural Network Accelerator Using Fine-Grained Logarithmic Quantization

被引:2
作者
Madadum, Hadee [1 ]
Becerikli, Yasar [1 ]
机构
[1] Kocaeli Univ, Dept Comp Engn, TR-41380 Kocaeli, Turkey
关键词
Convolutional neural network; logarithmic quantization; FPGA; resource efficiency;
D O I
10.32604/iasc.2022.023831
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional Neural Network (ConNN) implementations on Field Programmable Gate Array (FPGA) are being studied since the computational capabilities of FPGA have been improved recently. Model compression is required to enable ConNN deployment on resource-constrained FPGA devices. Logarithmic quantization is one of the efficient compression methods that can compress a model to very low bit-width without significant deterioration in performance. It is also hardware-friendly by using bitwise operations for multiplication. However, the logarithmic suffers from low resolution at high inputs due to exponential properties. Therefore, we propose a modified logarithmic quantization method with a fine resolution to compress a neural network model. In experiments, quantized models achieve a negligible loss of accuracy without the need for retraining steps. Besides this, we propose a resource-efficient hardware accelerator for running ConNN inference. Our design completely eliminates multipliers with bit shifters and adders. Throughput is measured in Giga Operation Per Second (GOP/s). The hardware utilization efficiency is represented by GOP/s per block of Digital Signal Processing (DSP) and Look-up Tables (LUTs). The result shows that the accelerator achieves resource efficiency of 9.38 GOP/s/DSP and 3.33 GOP/s/kLUTs.
引用
收藏
页码:681 / 695
页数:15
相关论文
共 19 条
[1]  
Banner R, 2019, ADV NEUR IN, V32
[2]   A Deep Look into Logarithmic Quantization of Model Parameters in Neural Networks [J].
Cai, Jingyong ;
Takemoto, Masashi ;
Nakajo, Hironori .
PROCEEDINGS OF THE 10TH INTERNATIONAL CONFERENCE ON ADVANCES IN INFORMATION TECHNOLOGY (IAIT2018), 2018,
[3]   ZeroQ: A Novel Zero Shot Quantization Framework [J].
Cai, Yaohui ;
Yao, Zhewei ;
Dong, Zhen ;
Gholami, Amir ;
Mahoney, Michael W. ;
Keutzer, Kurt .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :13166-13175
[4]  
Chang SE, 2021, INT S HIGH PERF COMP, P208, DOI [10.1109/HPCA51647.2021.00027, 10.1109/WRCSARA53879.2021.9612678]
[5]  
Choi J., 2018, IEEE C COMP VIS PATT
[6]  
Goyal R., CORR ABS210202147, P2021
[7]   MXQN:Mixed quantization for reducing bit-width of weights and activations in deep convolutional neural networks [J].
Huang, Chenglong ;
Liu, Puguang ;
Fang, Liang .
APPLIED INTELLIGENCE, 2021, 51 (07) :4561-4574
[8]  
Krishnamoorthi Raghuraman, 2018, ARXIV180608342V1
[9]  
Lee EH, 2017, INT CONF ACOUST SPEE, P5900, DOI 10.1109/ICASSP.2017.7953288
[10]  
Li Yunzhu, 2020, 8 INT C LEARNING REP