SpDRAM: Efficient In-DRAM Acceleration of Sparse Matrix-Vector Multiplication

被引:0
作者
Kang, Jieui [1 ]
Choi, Soeun [1 ]
Lee, Eunjin [1 ]
Sim, Jaehyeong [2 ]
机构
[1] Ewha Womans Univ, Artificial Intelligence Convergence, Seoul 03760, South Korea
[2] Ewha Womans Univ, Dept Comp Sci & Engn, Seoul 03760, South Korea
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Random access memory; Sparse matrices; Computer architecture; Logic; Vectors; Turning; System-on-chip; Space exploration; Sorting; SRAM cells; Processing-in-memory; SpMV; sparsity; DRAM; ARCHITECTURE;
D O I
10.1109/ACCESS.2024.3505622
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We introduce novel sparsity-aware in-DRAM matrix mapping techniques and a correspondingDRAM-based acceleration framework, termedSpDRAM, which utilizes a triple row activation schemeto efficiently handle sparse matrix-vector multiplication (SpMV). We found that reducing operationsby sparsity relies heavily on how matrices are mapped into DRAM banks, which operate row byrow. These banks operate row by row. From this insight, we developed two distinct matrix mappingtechniques aimed at maximizing the reduction of row operations with minimal design overhead: Output-aware Matrix Permutation (OMP) and Zero-aware Matrix Column Sorting (ZMCS). Additionally,we propose a Multiplication Deferring (MD) scheme that leverages the prevalent bit-level sparsity inmatrix values to decrease the effective bit-width required for in-bank multiplication operations. Evaluationresults demonstrate that the combination of our in-DRAM acceleration methods outperforms the latestDRAM-based PIM accelerator for SpMV, achieving a performance increase of up to 7.54xand a 22.4ximprovement in energy efficiency in a wide range of SpMV tasks
引用
收藏
页码:176009 / 176021
页数:13
相关论文
共 50 条
  • [21] Multiple-precision sparse matrix-vector multiplication on GPUs
    Isupov, Konstantin
    JOURNAL OF COMPUTATIONAL SCIENCE, 2022, 61
  • [22] A sparse matrix-vector multiplication method with low preprocessing cost
    Aktemur, Baris
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2018, 30 (21)
  • [23] Automatic tuning of sparse matrix-vector multiplication on multicore clusters
    LI ShiGang
    HU ChangJun
    ZHANG JunChao
    ZHANG YunQuan
    Science China(Information Sciences), 2015, 58 (09) : 17 - 30
  • [24] Iterative Sparse Matrix-Vector Multiplication for Integer Factorization on GPUs
    Schmidt, Bertil
    Aribowo, Hans
    Dang, Hoang-Vu
    EURO-PAR 2011 PARALLEL PROCESSING, PT 2, 2011, 6853 : 413 - 424
  • [25] TileSpMV: A Tiled Algorithm for Sparse Matrix-Vector Multiplication on GPUs
    Niu, Yuyao
    Lu, Zhengyang
    Dong, Meichen
    Jin, Zhou
    Liu, Weifeng
    Tan, Guangming
    2021 IEEE 35TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS), 2021, : 68 - 78
  • [26] Merge-based Parallel Sparse Matrix-Vector Multiplication
    Merrill, Duane
    Garland, Michael
    SC '16: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2016, : 678 - 689
  • [27] Automatic tuning of sparse matrix-vector multiplication on multicore clusters
    Li ShiGang
    Hu ChangJun
    Zhang JunChao
    Zhang YunQuan
    SCIENCE CHINA-INFORMATION SCIENCES, 2015, 58 (09) : 1 - 14
  • [28] Fast Sparse Matrix-Vector Multiplication on GPUs for Graph Applications
    Ashari, Arash
    Sedaghati, Naser
    Eisenlohr, John
    Parthasarathy, Srinivasan
    Sadayappan, P.
    SC14: INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2014, : 781 - 792
  • [29] Characterizing Dataset Dependence for Sparse Matrix-Vector Multiplication on GPUs
    Sedaghati, Naser
    Ashari, Arash
    Pouchet, Louis-Noel
    Parthasarathy, Srinivasan
    Sadayappan, P.
    2ND WORKSHOP ON PARALLEL PROGRAMMING FOR ANALYTICS APPLICATIONS (PPAA 2015), 2015, : 17 - 24
  • [30] TaiChi: A Hybrid Compression Format for Binary Sparse Matrix-Vector Multiplication on GPU
    Gao, Jianhua
    Ji, Weixing
    Tan, Zhaonian
    Wang, Yizhuo
    Shi, Feng
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2022, 33 (12) : 3732 - 3745