Reducing inter-process communication overhead in parallel sparse matrix-matrix multiplication

被引:1
|
作者
Ahmed M.S. [1 ]
Houser J. [1 ]
Hoque M.A. [1 ]
Raju R. [2 ]
Pfeiffer P. [1 ]
机构
[1] East Tennessee State University, Johnson City, TN
[2] University of Houston, Houston, TX
来源
Int. J. Grid High Perform. Comput. | / 3卷 / 46-59期
关键词
Communication Overhead; MPI Communication; Parallel Computing; Performance Analysis; Scalability; Sparse Matrix-Matrix Multiplication;
D O I
10.4018/IJGHPC.2017070104
中图分类号
学科分类号
摘要
Parallel sparse matrix-matrix multiplication algorithms (PSpGEMM) spend most of their running time on inter-process communication. In the case of distributed matrix-matrix multiplications, much of this time is spent on interchanging the partial results that are needed to calculate the final product matrix. This overhead can be reduced with a one-dimensional distributed algorithm for parallel sparse matrix-matrix multiplication that uses a novel accumulation pattern based on the logarithmic complexity of the number of processors (i.e., O (log (p)) where p is the number of processors). This algorithm's MPI communication overhead and execution time were evaluated on an HPC cluster, using randomly generated sparse matrices with dimensions up to one million by one million. The results showed a reduction of inter-process communication overhead for matrices with larger dimensions compared to another one dimensional parallel algorithm that takes O(p) run-time complexity for accumulating the results. Copyright © 2017, IGI Global.
引用
收藏
页码:46 / 59
页数:13
相关论文
共 50 条
  • [1] Communication-Avoiding Parallel Sparse-Dense Matrix-Matrix Multiplication
    Koanantakool, Penporn
    Azad, Ariful
    Buluc, Aydin
    Morozov, Dmitriy
    Oh, Sang-Yun
    Oliker, Leonid
    Yelick, Katherine
    2016 IEEE 30TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS 2016), 2016, : 842 - 853
  • [2] Parallel Efficient Sparse Matrix-Matrix Multiplication on Multicore Platforms
    Patwary, Md. Mostofa Ali
    Satish, Nadathur Rajagopalan
    Sundaram, Narayanan
    Park, Jongsoo
    Anderson, Michael J.
    Vadlamudi, Satya Gautam
    Das, Dipankar
    Pudov, Sergey G.
    Pirogov, Vadim O.
    Dubey, Pradeep
    HIGH PERFORMANCE COMPUTING, ISC HIGH PERFORMANCE 2015, 2015, 9137 : 48 - 57
  • [3] PARALLEL SPARSE MATRIX-MATRIX MULTIPLICATION AND INDEXING: IMPLEMENTATION AND EXPERIMENTS
    Buluc, Aydin
    Gilbert, John R.
    SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2012, 34 (04): : C170 - C191
  • [4] Partitioning Models for Scaling Parallel Sparse Matrix-Matrix Multiplication
    Akbudak, Kadir
    Selvitopi, Oguz
    Aykanat, Cevdet
    ACM TRANSACTIONS ON PARALLEL COMPUTING, 2018, 4 (03)
  • [5] Register-Aware Optimizations for Parallel Sparse Matrix-Matrix Multiplication
    Liu, Junhong
    He, Xin
    Liu, Weifeng
    Tan, Guangming
    INTERNATIONAL JOURNAL OF PARALLEL PROGRAMMING, 2019, 47 (03) : 403 - 417
  • [6] Brief Announcement: Hypergraph Partitioning for Parallel Sparse Matrix-Matrix Multiplication
    Ballard, Grey
    Druinsky, Alex
    Knight, Nicholas
    Schwartz, Oded
    SPAA'15: PROCEEDINGS OF THE 27TH ACM SYMPOSIUM ON PARALLELISM IN ALGORITHMS AND ARCHITECTURES, 2015, : 86 - 88
  • [7] PERFORMANCE EVALUATION OF SPARSE MATRIX-MATRIX MULTIPLICATION
    Jain-Mendon, Shweta
    Sass, Ron
    2013 23RD INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE LOGIC AND APPLICATIONS (FPL 2013) PROCEEDINGS, 2013,
  • [8] Hypergraph partitioning for sparse matrix-matrix multiplication
    Ballard G.
    Druinsky A.
    Knight N.
    Schwartz O.
    ACM Transactions on Parallel Computing, 2016, 3 (03) : 1 - 34
  • [9] Optimizing Sparse Matrix-Matrix Multiplication for the GPU
    Dalton, Steven
    Olson, Luke
    Bell, Nathan
    ACM TRANSACTIONS ON MATHEMATICAL SOFTWARE, 2015, 41 (04):
  • [10] Sparse Matrix-Matrix Multiplication on Modern Architectures
    Matam, Kiran
    Indarapu, Siva Rama Krishna Bharadwaj
    Kothapalli, Kishore
    2012 19TH INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPUTING (HIPC), 2012,