RISCV-FNT: A Fast FNT-based RISC-V Processor for CNN Acceleration

被引:0
作者
Chen, Bingzhen [1 ]
Wang, Xingbo [1 ]
Huang, Yucong [1 ,2 ]
Xu, Zhiyuan [1 ]
机构
[1] Care of Ye TT, Southern Univ Sci & Technol, Shenzhen, Peoples R China
[2] Hong Kong Univ Sci & Technol, Hong Kong, Peoples R China
来源
2024 IEEE 6TH INTERNATIONAL CONFERENCE ON AI CIRCUITS AND SYSTEMS, AICAS 2024 | 2024年
关键词
Convolutional Neural Network; Fermat Number Transform; RISC-V; Custom Instruction;
D O I
10.1109/AICAS59952.2024.10595907
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolution forms the basis of computation in neural network applications. Many different approaches have been proposed in the past years to optimize the convolution operation. In this paper, we propose to use Fermat Number Transform (FNT) technique to accelerate the computation of convolution in neural networks. Calculations in FNT are all based on real numbers, which significantly reduce the complexity as compared to complex-number-based FFT calculations. Furthermore, by using diminished-1 encoding, multiplication and modulo operations can also be simplified into bit manipulations. In this paper, we have constructed a RISC-V based processor, called RISCV-FNT, which incorporates an FNT-based convolution acceleration unit, along with custom instruction sets. FPGA implementation of RISCV-FNT demonstrated an 8.5x speedup compared to other RISC-V processors without FNT acceleration when performing inference tasks on Lenet-5. Synthesized results from Synopsys (R) DC achieved area energy efficiency of 93.9 GOPs/W/mm(2).
引用
收藏
页码:292 / 296
页数:5
相关论文
共 11 条
  • [1] Chen YH, 2016, ISSCC DIG TECH PAP I, V59, P262, DOI 10.1109/ISSCC.2016.7418007
  • [2] Cong Jason, 2014, Artificial Neural Networks and Machine Learning - ICANN 2014. 24th International Conference on Artificial Neural Networks. Proceedings: LNCS 8681, P281, DOI 10.1007/978-3-319-11179-7_36
  • [3] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [4] Fast Algorithms for Convolutional Neural Networks
    Lavin, Andrew
    Gray, Scott
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 4013 - 4021
  • [5] SIMPLIFIED BINARY ARITHMETIC FOR FERMAT NUMBER TRANSFORM
    LEIBOWITZ, LM
    [J]. IEEE TRANSACTIONS ON ACOUSTICS SPEECH AND SIGNAL PROCESSING, 1976, 24 (05): : 356 - 359
  • [6] Nguyen-Thanh N, 2016, PROC INT CONF ADV, P231, DOI 10.1109/ATC.2016.7764779
  • [7] You Only Look Once: Unified, Real-Time Object Detection
    Redmon, Joseph
    Divvala, Santosh
    Girshick, Ross
    Farhadi, Ali
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 779 - 788
  • [8] Simonyan K, 2015, Arxiv, DOI arXiv:1409.1556
  • [9] Deep Convolutional Neural Network Architecture With Reconfigurable Computation Patterns
    Tu, Fengbin
    Yin, Shouyi
    Ouyang, Peng
    Tang, Shibin
    Liu, Leibo
    Wei, Shaojun
    [J]. IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2017, 25 (08) : 2220 - 2233
  • [10] Reconfigurable and Low-Complexity Accelerator for Convolutional and Generative Networks Over Finite Fields
    Xu, Weihong
    Zhang, Zaichen
    You, Xiaohu
    Zhang, Chuan
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2020, 39 (12) : 4894 - 4907