The Accuracy and Efficiency of Posit Arithmetic

被引:3
作者
Ciocirlan, Stefan Dan [1 ,2 ]
Loghin, Dumitrel [1 ]
Ramapantulu, Lavanya
Tapus, Nicolae [2 ]
Teo, Yong Meng [1 ]
机构
[1] Natl Univ Singapore, Dept Comp Sci, Singapore, Singapore
[2] Univ Politehn Bucuresti, Dept Comp Sci, Bucharest, Romania
来源
2021 IEEE 39TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2021) | 2021年
关键词
posit; floating-point; IEEE; 754; RISC-V; accuracy; efficiency;
D O I
10.1109/ICCD53106.2021.00024
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Motivated by the increasing interest in the posit numeric format, in this paper we evaluate the accuracy and efficiency of posit arithmetic in contrast to the traditional IEEE 754 32-bit floating-point (FP32) arithmetic. We first design and implement a Posit Arithmetic Unit (PAU), called POSAR, with flexible bit-sized arithmetic suitable for applications that can trade accuracy for savings in chip area. Next, we analyze the accuracy and efficiency of POSAR with a series of benchmarks including mathematical computations, ML kernels, NAS Parallel Benchmarks (NPB), and Cifar-10 CNN. This analysis is done on our implementation of POSAR integrated into a RISC-V Rocket Chip core in comparison with the IEEE 754-based Floting Point Unit (FPU) of Rocket Chip. Our analysis shows that POSAR can outperform the FPU, but the results are not spectacular. For NPB, 32-bit posit achieves better accuracy than FP32 and improves the execution by up to 2%. However, POSAR with 32-bit posit needs 30% more FPGA resources compared to the FPU. For classic ML algorithms, we find that 8-bit posits are not suitable to replace FP32 because they exhibit low accuracy leading to wrong results. Instead, 16-bit posit offers the best option in terms of accuracy and efficiency. For example, 16-bit posit achieves the same Ibp-1 accuracy as FP32 on a Cifar-10 CNN with a speedup of 18%.
引用
收藏
页码:83 / 87
页数:5
相关论文
共 24 条
  • [1] Asanovic K, 2016, UCBEECS201617
  • [2] Bailey D. H., 2011, Encyclopedia of Parallel Computing, P1254
  • [3] SMURF: Scalar Multiple-precision Unum Risc-V Floating-point Accelerator for Scientific Computing
    Bocco, Andrea
    Durand, Yves
    De Dinechin, Florent
    [J]. CONFERENCE FOR NEXT GENERATION ARITHMETIC 2019 (CONGA), 2019,
  • [4] Performance-Efficiency Trade-off of Low-Precision Numerical Formats in Deep Neural Networks
    Carmichael, Zachariah
    Langroudi, Hamed F.
    Khazanov, Char
    Lillie, Jeffrey
    Gustafson, John L.
    Kudithipudi, Dhireesha
    [J]. CONFERENCE FOR NEXT GENERATION ARITHMETIC 2019 (CONGA), 2019,
  • [5] Parameterized Posit Arithmetic Hardware Generator
    Chaurasiya, Rohit
    Gustafson, John
    Shrestha, Rahul
    Neudorfer, Jonathan
    Nambiar, Sangeeth
    Niyogi, Kaustav
    Merchant, Farhad
    Leupers, Rainer
    [J]. 2018 IEEE 36TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD), 2018, : 334 - 341
  • [6] Chien Steven W. D., 2020, Parallel Processing and Applied Mathematics. 13th International Conference, PPAM 2019. Revised Selected Papers. Lecture Notes in Computer Science (LNCS 12043), P301, DOI 10.1007/978-3-030-43229-4_26
  • [7] Posits: the good, the bad and the ugly
    de Dinechin, Florent
    Forget, Luc
    Muller, Jean-Michel
    Uguen, Yohann
    [J]. CONFERENCE FOR NEXT GENERATION ARITHMETIC 2019 (CONGA), 2019,
  • [8] AN ASIC PERSPECTIVE ON FPGA OPTIMIZATIONS
    Ehliar, Andreas
    Liu, Dake
    [J]. FPL: 2009 INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE LOGIC AND APPLICATIONS, 2009, : 218 - 223
  • [9] Tutorial: SHAKTI Processors: An Open-Source Hardware Initiative
    Gala, Neel
    Menon, Arjun
    Bodduna, Rahul
    Madhusudan, G. S.
    Kamakoti, V.
    [J]. 2016 29TH INTERNATIONAL CONFERENCE ON VLSI DESIGN AND 2016 15TH INTERNATIONAL CONFERENCE ON EMBEDDED SYSTEMS (VLSID), 2016, : 7 - 8
  • [10] Low Complexity Multiply-Accumulate Units for Convolutional Neural Networks with Weight-Sharing
    Garland, James
    Gregg, David
    [J]. ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2018, 15 (03)