Reducing inter-process communication overhead in parallel sparse matrix-matrix multiplication

被引:1
|
作者
Ahmed M.S. [1 ]
Houser J. [1 ]
Hoque M.A. [1 ]
Raju R. [2 ]
Pfeiffer P. [1 ]
机构
[1] East Tennessee State University, Johnson City, TN
[2] University of Houston, Houston, TX
来源
Int. J. Grid High Perform. Comput. | / 3卷 / 46-59期
关键词
Communication Overhead; MPI Communication; Parallel Computing; Performance Analysis; Scalability; Sparse Matrix-Matrix Multiplication;
D O I
10.4018/IJGHPC.2017070104
中图分类号
学科分类号
摘要
Parallel sparse matrix-matrix multiplication algorithms (PSpGEMM) spend most of their running time on inter-process communication. In the case of distributed matrix-matrix multiplications, much of this time is spent on interchanging the partial results that are needed to calculate the final product matrix. This overhead can be reduced with a one-dimensional distributed algorithm for parallel sparse matrix-matrix multiplication that uses a novel accumulation pattern based on the logarithmic complexity of the number of processors (i.e., O (log (p)) where p is the number of processors). This algorithm's MPI communication overhead and execution time were evaluated on an HPC cluster, using randomly generated sparse matrices with dimensions up to one million by one million. The results showed a reduction of inter-process communication overhead for matrices with larger dimensions compared to another one dimensional parallel algorithm that takes O(p) run-time complexity for accumulating the results. Copyright © 2017, IGI Global.
引用
收藏
页码:46 / 59
页数:13
相关论文
共 50 条
  • [21] Computationally efficient parallel matrix-matrix multiplication on the torus
    Zekri, Ahmed S.
    Sedukhin, Stanislav G.
    HIGH-PERFORMANCE COMPUTING, 2008, 4759 : 219 - 226
  • [22] SIMULTANEOUS INPUT AND OUTPUT MATRIX PARTITIONING FOR OUTER-PRODUCT-PARALLEL SPARSE MATRIX-MATRIX MULTIPLICATION
    Akbudak, Kadir
    Aykanat, Cevdet
    SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2014, 36 (05): : C568 - C590
  • [23] Parallel photonic acceleration processor for matrix-matrix multiplication
    Huang, Ying
    Yue, Hengsong
    Ma, Wei
    Zhang, Yiyuan
    Xiao, Yao
    Tang, Yong
    Tang, He
    Chu, Tao
    OPTICS LETTERS, 2023, 48 (12) : 3231 - 3234
  • [24] Design space exploration for sparse matrix-matrix multiplication on FPGAs
    Lin, Colin Yu
    Wong, Ngai
    So, Hayden Kwok-Hay
    INTERNATIONAL JOURNAL OF CIRCUIT THEORY AND APPLICATIONS, 2013, 41 (02) : 205 - 219
  • [25] EXPLOITING MULTIPLE LEVELS OF PARALLELISM IN SPARSE MATRIX-MATRIX MULTIPLICATION
    Azad, Ariful
    Ballard, Grey
    Buluc, Aydin
    Demmel, James
    Grigori, Laura
    Schwartz, Oded
    Toledo, Sivan
    Williams, Samuel
    SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2016, 38 (06): : C624 - C651
  • [26] Accelerating sparse matrix-matrix multiplication with GPU Tensor Cores
    Zachariadis, Orestis
    Satpute, Nitin
    Gomez-Luna, Juan
    Olivares, Joaquin
    COMPUTERS & ELECTRICAL ENGINEERING, 2020, 88 (88)
  • [27] Communication-Avoiding and Memory-Constrained Sparse Matrix-Matrix Multiplication at Extreme Scale
    Hussain, Md Taufique
    Selvitopi, Oguz
    Buluc, Aydin
    Azad, Ariful
    2021 IEEE 35TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS), 2021, : 90 - 100
  • [28] Parallel Algorithms for Masked Sparse Matrix-Matrix Products
    Milakovic, Srdan
    Selvitopi, Oguz
    Nisa, Israt
    Budimlic, Zoran
    Buluc, Aydin
    51ST INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, ICPP 2022, 2022,
  • [29] Sparse approximate matrix-matrix multiplication for density matrix purification with error control
    Artemov, Anton G.
    Rubensson, Emanuel H.
    JOURNAL OF COMPUTATIONAL PHYSICS, 2021, 438
  • [30] Parallel Algorithm for Quasi-Band Matrix-Matrix Multiplication
    Vooturi, Dharma Teja
    Kothapalli, Kishore
    PARALLEL PROCESSING AND APPLIED MATHEMATICS, PPAM 2015, PT I, 2016, 9573 : 106 - 115