SpDRAM: Efficient In-DRAM Acceleration of Sparse Matrix-Vector Multiplication

被引:0
|
作者
Kang, Jieui [1 ]
Choi, Soeun [1 ]
Lee, Eunjin [1 ]
Sim, Jaehyeong [2 ]
机构
[1] Ewha Womans Univ, Artificial Intelligence Convergence, Seoul 03760, South Korea
[2] Ewha Womans Univ, Dept Comp Sci & Engn, Seoul 03760, South Korea
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Random access memory; Sparse matrices; Computer architecture; Logic; Vectors; Turning; System-on-chip; Space exploration; Sorting; SRAM cells; Processing-in-memory; SpMV; sparsity; DRAM; ARCHITECTURE;
D O I
10.1109/ACCESS.2024.3505622
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We introduce novel sparsity-aware in-DRAM matrix mapping techniques and a correspondingDRAM-based acceleration framework, termedSpDRAM, which utilizes a triple row activation schemeto efficiently handle sparse matrix-vector multiplication (SpMV). We found that reducing operationsby sparsity relies heavily on how matrices are mapped into DRAM banks, which operate row byrow. These banks operate row by row. From this insight, we developed two distinct matrix mappingtechniques aimed at maximizing the reduction of row operations with minimal design overhead: Output-aware Matrix Permutation (OMP) and Zero-aware Matrix Column Sorting (ZMCS). Additionally,we propose a Multiplication Deferring (MD) scheme that leverages the prevalent bit-level sparsity inmatrix values to decrease the effective bit-width required for in-bank multiplication operations. Evaluationresults demonstrate that the combination of our in-DRAM acceleration methods outperforms the latestDRAM-based PIM accelerator for SpMV, achieving a performance increase of up to 7.54xand a 22.4ximprovement in energy efficiency in a wide range of SpMV tasks
引用
收藏
页码:176009 / 176021
页数:13
相关论文
共 50 条
  • [1] MViD: Sparse Matrix-Vector Multiplication in Mobile DRAM for Accelerating Recurrent Neural Networks
    Kim, Byeongho
    Chung, Jongwook
    Lee, Eojin
    Jung, Wonkyung
    Lee, Sunjung
    Choi, Jaewan
    Park, Jaehyun
    Wi, Minbok
    Lee, Sukhan
    Ahn, Jung Ho
    IEEE TRANSACTIONS ON COMPUTERS, 2020, 69 (07) : 955 - 967
  • [2] Sparse Matrix-Vector Multiplication on GPGPUs
    Filippone, Salvatore
    Cardellini, Valeria
    Barbieri, Davide
    Fanfarillo, Alessandro
    ACM TRANSACTIONS ON MATHEMATICAL SOFTWARE, 2017, 43 (04):
  • [3] Adaptive sparse matrix representation for efficient matrix-vector multiplication
    Zardoshti, Pantea
    Khunjush, Farshad
    Sarbazi-Azad, Hamid
    JOURNAL OF SUPERCOMPUTING, 2016, 72 (09) : 3366 - 3386
  • [4] Processor-efficient sparse matrix-vector multiplication
    Heath, LS
    Ribbens, CJ
    Pemmaraju, SV
    COMPUTERS & MATHEMATICS WITH APPLICATIONS, 2004, 48 (3-4) : 589 - 608
  • [5] Efficient Sparse Matrix-Vector Multiplication on Intel PIUMA Architecture
    Aananthakrishnan, Sriram
    Pawlowski, Robert
    Fryman, Joshua
    Hur, Ibrahim
    2020 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2020,
  • [6] An efficient SIMD compression format for sparse matrix-vector multiplication
    Chen, Xinhai
    Xie, Peizhen
    Chi, Lihua
    Liu, Jie
    Gong, Chunye
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2018, 30 (23)
  • [7] Efficient Multicore Sparse Matrix-Vector Multiplication for FE Electromagnetics
    Fernandez, David M.
    Giannacopoulos, Dennis
    Gross, Warren J.
    IEEE TRANSACTIONS ON MAGNETICS, 2009, 45 (03) : 1392 - 1395
  • [8] Sparse Matrix-Vector Multiplication Based on Online Arithmetic
    Cherati, Sahar Moradi
    Jaberipur, Ghassem
    Sousa, Leonel
    IEEE ACCESS, 2024, 12 : 87653 - 87664
  • [9] Node aware sparse matrix-vector multiplication
    Bienz, Amanda
    Gropp, William D.
    Olson, Luke N.
    JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 2019, 130 : 166 - 178
  • [10] An Efficient Sparse Matrix-Vector Multiplication on Distributed Memory Parallel Computers
    Shahnaz, Rukhsana
    Usman, Anila
    INTERNATIONAL JOURNAL OF COMPUTER SCIENCE AND NETWORK SECURITY, 2007, 7 (01): : 77 - 82