DSAV: A Deep Sparse Acceleration Framework for Voxel-Based 3-D Object Detection

被引:0
作者
Fang, Haining [1 ]
Tan, Yujuan [1 ]
Ren, Ao [1 ]
Zhuang, Wei [2 ]
Hua, Yang
Qin, Zhiyong
Liu, Duo [3 ]
机构
[1] Chongqing Univ, Coll Comp Sci, Chongqing 400044, Peoples R China
[2] Beijing Microelect Technol Inst, SOC Dept, Beijing 100076, Peoples R China
[3] Chongqing Univ, Sch Big Data & Software Engn, Chongqing 400044, Peoples R China
关键词
Three-dimensional displays; Convolution; Object detection; Solid modeling; Hardware; Computational modeling; Point cloud compression; 3-D object detection; accelerator; structured pruning; voxelization;
D O I
10.1109/TCAD.2024.3437334
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Voxel-based 3-D object detection has been widely applied in robotics, virtual reality, and autonomous driving. However, inefficiency in the voxelization and backbone-network computation, which are the main components of the voxel-based models, prevents efficient 3-D object detection. First, due to the high sparsity and irregularity of the point cloud, the voxelization process usually requires generalized platforms, such as CPUs, and causes low voxelization speed. Second, the voxel-based models contain considerable transposed convolutional layers, and existing accelerators introduce considerable additional hardware to support both the convolution and transposed convolution operations. Nonetheless, this strategy incurs significant hardware costs. Besides, transposed convolutions result in various patterns of sparse feature maps, and pruning as a representative model compression technique, results in sparse weight matrices. The two types of sparsity impose challenges in accelerating the voxel-based models, including activation-weight matching efficiency, low partial-sum accumulation efficiency, and workload imbalance issues. In this work, we propose DSAV, a 3-D object detection accelerator to address these obstacles. Specifically, we first propose a hash-based voxelizer for efficient voxelization, by storing and indexing voxels hierarchically. Then, we collaboratively design the transposed convolution acceleration method, structured pruning method, and accelerator architecture for the voxel-based models. As a result, the accelerator can fully leverage the sparsity lies in both feature maps and weight matrices. Experimental results show that the proposed accelerator can outperform the prior studies by 19x similar to 19.8x faster in voxelization and 4.29x similar to 38.01x faster in backbone inference. Finally, the accelerator achieves 4.61x similar to 31.63x speedups than its counterparts in 3-D object detection tasks.
引用
收藏
页码:613 / 626
页数:14
相关论文
共 36 条
  • [1] [Anonymous], 2022, Vitis-AI
  • [2] Bansal M, 2011, IEEE INT CONF ROBOT, P1803
  • [3] FINN-R: An End-to-End Deep-Learning Framework for Fast Exploration of Quantized Neural Networks
    Blott, Michaela
    Preusser, Thomas B.
    Fraser, Nicholas J.
    Gambardella, Giulio
    O'Brien, Kenneth
    Umuroglu, Yaman
    Leeser, Miriam
    Vissers, Kees
    [J]. ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2018, 11 (03)
  • [4] CARTER JL, 1979, J COMPUT SYST SCI, V18, P143, DOI 10.1016/0022-0000(79)90044-8
  • [5] An Energy-Efficient FPGA-Based Deconvolutional Neural Networks Accelerator for Single Image Super-Resolution
    Chang, Jung-Woo
    Kang, Keon-Woo
    Kang, Suk-Ju
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (01) : 281 - 295
  • [6] Chen Xizi, 2020, DESIGN AUTOMATION C
  • [7] Chen YH, 2016, ISSCC DIG TECH PAP I, V59, P262, DOI 10.1109/ISSCC.2016.7418007
  • [8] Cho JM, 2014, 2014 INTERNATIONAL SYMPOSIUM ON VLSI DESIGN, AUTOMATION AND TEST (VLSI-DAT)
  • [9] Deng JJ, 2021, AAAI CONF ARTIF INTE, V35, P1201
  • [10] RTMobile: Beyond Real-Time Mobile Acceleration of RNNs for Speech Recognition
    Dong, Peiyan
    Wang, Siyue
    Niu, Wei
    Zhang, Chengming
    Lin, Sheng
    Li, Zhengang
    Gong, Yifan
    Ren, Bin
    Lin, Xue
    Tao, Dingwen
    [J]. PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2020,