Variable-Precision Approximate Floating-Point Multiplier for Efficient Deep Learning Computation

被引:9
作者
Zhang, Hao [1 ]
Ko, Seok-Bum [2 ]
机构
[1] Ocean Univ China, Fac Informat Sci & Engn, Qingdao 266100, Peoples R China
[2] Univ Saskatchewan, Dept Elect & Comp Engn, Saskatoon, SK S7N 5A9, Canada
关键词
Deep learning; Encoding; Computer architecture; Computational efficiency; Circuits and systems; Adders; Hardware design languages; Approximate multiplier; posit format; deep learning computation; variable precision;
D O I
10.1109/TCSII.2022.3161005
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this brief, a variable-precision approximate floating-point multiplier is proposed for energy efficient deep learning computation. The proposed architecture supports approximate multiplication with BFloat16 format. As the input and output activations of deep learning models usually follow normal distribution, inspired by the posit format, for numbers with different values, different precisions can be applied to represent them. In the proposed architecture, posit encoding is used to change the level of approximation, and the precision of the computation is controlled by the value of product exponent. For large exponent, smaller precision multiplication is applied to mantissa and for small exponent, higher precision computation is applied. Truncation is used as approximate method in the proposed design while the number of bit positions to be truncated is controlled by the values of the product exponent. The proposed design can achieve 19% area reduction and 42% power reduction compared to the normal BFloat16 multiplier. When applying the proposed multiplier in deep learning computation, almost the same accuracy as that of normal BFloat16 multiplier can be achieved.
引用
收藏
页码:2503 / 2507
页数:5
相关论文
共 25 条
  • [1] DLFloat: A 16-b Floating Point format designed for Deep Learning Training and Inference
    Agrawal, Ankur
    Mueller, Silvia M.
    Fleischer, Bruce M.
    Choi, Jungwook
    Wang, Naigang
    Sun, Xiao
    Gopalakrishnan, Kailash
    [J]. 2019 IEEE 26TH SYMPOSIUM ON COMPUTER ARITHMETIC (ARITH), 2019, : 92 - 95
  • [2] [Anonymous], 2017, Deep convolutional neural network inference with floating-point weights and fixed-point activations
  • [3] Improving the Accuracy and Hardware Efficiency of Neural Networks Using Approximate Multipliers
    Ansari, Mohammad Saeed
    Mrazek, Vojtech
    Cockburn, Bruce F.
    Sekanina, Lukas
    Vasicek, Zdenek
    Han, Jie
    [J]. IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2020, 28 (02) : 317 - 328
  • [4] A SIGNED BINARY MULTIPLICATION TECHNIQUE
    BOOTH, AD
    [J]. QUARTERLY JOURNAL OF MECHANICS AND APPLIED MATHEMATICS, 1951, 4 (02) : 236 - 240
  • [5] Camus V, 2016, PROC EUR SOLID-STATE, P465, DOI 10.1109/ESSCIRC.2016.7598342
  • [6] Chen CY, 2018, DES AUT TEST EUROPE, P821, DOI 10.23919/DATE.2018.8342119
  • [7] Gustafson John L., 2017, [Supercomputing Frontiers and Innovations, Supercomputing Frontiers and Innovations], V4, P71
  • [8] Han, 2013, TEST S ETS 2013 18 I, P1, DOI DOI 10.1109/ETS.2013.6569370
  • [9] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [10] Leveraging the bfloat16 Artificial Intelligence Datatype For Higher-Precision Computations
    Henry, Greg
    Tang, Ping Tak Peter
    Heinecke, Alexander
    [J]. 2019 IEEE 26TH SYMPOSIUM ON COMPUTER ARITHMETIC (ARITH), 2019, : 69 - 76