Reducing inter-process communication overhead in parallel sparse matrix-matrix multiplication

被引:1
|
作者
Ahmed M.S. [1 ]
Houser J. [1 ]
Hoque M.A. [1 ]
Raju R. [2 ]
Pfeiffer P. [1 ]
机构
[1] East Tennessee State University, Johnson City, TN
[2] University of Houston, Houston, TX
来源
Int. J. Grid High Perform. Comput. | / 3卷 / 46-59期
关键词
Communication Overhead; MPI Communication; Parallel Computing; Performance Analysis; Scalability; Sparse Matrix-Matrix Multiplication;
D O I
10.4018/IJGHPC.2017070104
中图分类号
学科分类号
摘要
Parallel sparse matrix-matrix multiplication algorithms (PSpGEMM) spend most of their running time on inter-process communication. In the case of distributed matrix-matrix multiplications, much of this time is spent on interchanging the partial results that are needed to calculate the final product matrix. This overhead can be reduced with a one-dimensional distributed algorithm for parallel sparse matrix-matrix multiplication that uses a novel accumulation pattern based on the logarithmic complexity of the number of processors (i.e., O (log (p)) where p is the number of processors). This algorithm's MPI communication overhead and execution time were evaluated on an HPC cluster, using randomly generated sparse matrices with dimensions up to one million by one million. The results showed a reduction of inter-process communication overhead for matrices with larger dimensions compared to another one dimensional parallel algorithm that takes O(p) run-time complexity for accumulating the results. Copyright © 2017, IGI Global.
引用
收藏
页码:46 / 59
页数:13
相关论文
共 50 条
  • [41] Exploiting Locality in Sparse Matrix-Matrix Multiplication on Many-Core Architectures
    Akbudak, Kadir
    Aykanat, Cevdet
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2017, 28 (08) : 2258 - 2271
  • [42] GPU-ACCELERATED SPARSE MATRIX-MATRIX MULTIPLICATION BY ITERATIVE ROW MERGING
    Gremse, Felix
    Hoefter, Andreas
    Schwen, Lars Ole
    Kiessling, Fabian
    Naumann, Uwe
    SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2015, 37 (01): : C54 - C71
  • [43] spECK: Accelerating GPU Sparse Matrix-Matrix Multiplication through Lightweight Analysis
    Parger, Mathias
    Winter, Martin
    Mlakar, Daniel
    Steinberger, Markus
    PROCEEDINGS OF THE 25TH ACM SIGPLAN SYMPOSIUM ON PRINCIPLES AND PRACTICE OF PARALLEL PROGRAMMING (PPOPP '20), 2020, : 362 - 375
  • [44] Matrix-matrix multiplication on heterogeneous platforms
    Beaumont, O
    Boudet, V
    Rastello, F
    Robert, Y
    2000 INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, PROCEEDINGS, 2000, : 289 - 298
  • [45] Efficient Sparse-Dense Matrix-Matrix Multiplication on GPUs Using the Customized Sparse Storage Format
    Shi, Shaohuai
    Wang, Qiang
    Chu, Xiaowen
    2020 IEEE 26TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS), 2020, : 19 - 26
  • [46] Parallel device for sparse matrix multiplication
    Vyzhikovski, R.
    Kanevskii, Yu.S.
    Maslennikov, O.V.
    Engineering Simulation, 1993, 11 (03): : 412 - 422
  • [47] IA-SpGEMM An Input-aware Auto-tuning Framework for Parallel Sparse Matrix-Matrix Multiplication
    Xie, Zhen
    Tan, Guangming
    Liu, Weifeng
    Sun, Ninghui
    INTERNATIONAL CONFERENCE ON SUPERCOMPUTING (ICS 2019), 2019, : 94 - 105
  • [48] Locality-aware parallel block-sparse matrix-matrix multiplication using the Chunks and Tasks programming model
    Rubensson, Emanuel H.
    Rudberg, Elias
    PARALLEL COMPUTING, 2016, 57 : 87 - 106
  • [49] REDUCING COMMUNICATION COSTS FOR SPARSE MATRIX MULTIPLICATION WITHIN ALGEBRAIC MULTIGRID
    Rallard, Grey
    Siefer, Christopher
    Hu, Jonathan
    SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2016, 38 (03): : C203 - C231
  • [50] An Efficient Gustavson-Based Sparse Matrix-Matrix Multiplication Accelerator on Embedded FPGAs
    Li, Shiqing
    Huai, Shuo
    Liu, Weichen
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (12) : 4671 - 4680