DBPS: Dynamic Block Size and Precision Scaling for Efficient DNN Training Supported by RISC-V ISA Extensions

被引:1
|
作者
Lee, Seunghyun [1 ]
Choi, Jeik [2 ]
Noh, Seockhwan [2 ]
Koo, Jahyun [2 ]
Kung, Jaeha [1 ]
机构
[1] Korea Univ, Sch Elect Engn, Seoul, South Korea
[2] DGIST, Dept EECS, Daegu, South Korea
关键词
Block floating point; dynamic control; hardware accelerator; neural network training; RISC-V custom instruction;
D O I
10.1109/DAC56929.2023.10248013
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Over the past decade, it has been found that deep neural networks (DNNs) perform better on visual perception and language understanding tasks as their size increases. However, this comes at the cost of high energy consumption and large memory requirement to train such large models. As the training DNNs necessitates a wide dynamic range in representing tensors, floating point formats are normally used. In this work, we utilize a block floating point (BFP) format that significantly reduces the size of tensors and the power consumption of arithmetic units. Unfortunately, prior work on BFP-based DNN training empirically selects the block size and the precision that maintain the training accuracy. To make the BFP-based training more feasible, we propose dynamic block size and precision scaling (DBPS) for highly efficient DNN training. We also present a hardware accelerator, called DBPS core, which supports the DBPS control by configuring arithmetic units with custom instructions extended in a RISC-V processor. As a result, the training time and energy consumption reduce by 67.1% and 72.0%, respectively, without hurting the training accuracy.
引用
收藏
页数:6
相关论文
共 4 条
  • [1] MiniFloats on RISC-V Cores: ISA Extensions With Mixed-Precision Short Dot Products
    Bertaccini, Luca
    Paulin, Gianna
    Cavalcante, Matheus
    Fischer, Tim
    Mach, Stefan
    Benini, Luca
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2024, 12 (04) : 1040 - 1055
  • [2] A Scalable RISC-V Vector Processor Enabling Efficient Multi-Precision DNN Inference
    Wang, Chuanning
    Fang, Chao
    Wu, Xiao
    Wang, Zhongfeng
    Lin, Jun
    2024 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS 2024, 2024,
  • [3] MiniFloat-NN and ExSdotp: An ISA Extension and a Modular Open Hardware Unit for Low-Precision Training on RISC-V Cores
    Bertaccini, Luca
    Paulin, Gianna
    Fischer, Tim
    Mach, Stefan
    Benini, Luca
    2022 IEEE 29TH SYMPOSIUM ON COMPUTER ARITHMETIC (ARITH 2022), 2022, : 1 - 8
  • [4] Mix-GEMM: Extending RISC-V CPUs for Energy-Efficient Mixed-Precision DNN Inference Using Binary Segmentation
    Fornt, Jordi
    Reggiani, Enrico
    Fontova-Muste, Pau
    Rodas, Narcis
    Pappalardo, Alessandro
    Unsal, Osman Sabri
    Kestelman, Adrian Cristal
    Altet, Josep
    Moll, Francesc
    Abella, Jaume
    IEEE TRANSACTIONS ON COMPUTERS, 2025, 74 (02) : 582 - 596