A Sparse Matrix Optimization Method for Graph Neural Networks Training

被引:0
作者
Yao, Tiechui [1 ,2 ]
Wang, Jue [1 ,2 ]
Gu, Junyu [1 ,2 ]
Shi, Yumeng [1 ,2 ]
Liu, Fang [1 ,2 ]
Wang, Xiaoguang [2 ]
Wang, Yangang [1 ,2 ]
Chi, Xuebin [1 ,2 ]
机构
[1] Chinese Acad Sci, Comp Network Informat Ctr, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
来源
KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT I, KSEM 2023 | 2023年 / 14117卷
基金
国家重点研发计划;
关键词
Sparse matrix format; Sparse matrix-vector multiplication; Performance model; Graph neural networks;
D O I
10.1007/978-3-031-40283-8_11
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph neural networks (GNN) have shown great application potential in scientific research applications, biomedicine, and other fields, which exhibit superior feature representation capabilities for graph data with non-Euclidean structures. These capabilities are enabled efficiently by sparse matrix-matrix multiplication (SPMM) and sparse matrix-vector multiplication (SPMV) that operate on sparse matrix representations of graph structures. However, SpMM has the characteristics of high memory occupation and irregular memory access, which leads to low storage and computational efficiency. To address the above issues, this paper proposes a sparse matrix optimization method, including a sparse matrix format and a performance model. The format, namely BMCOO, divides the sparse matrix into multiple blocks and adopts the bitmap to compress the position information of non-zero elements in each block. This paper further designs an SpMV algorithm in BMCOO format on GPU. In addition, a multi-channel SpMV performance model is constructed to predict the execution time of SpMV by combining the sparse matrix scale and system architecture parameters. Then the performance model fine-tunes the graph partitioning of the GNN training process. Experiments on the SuiteSparse and the Open Graph Benchmark datasets verify the effectiveness and superiority of the proposed method.
引用
收藏
页码:114 / 123
页数:10
相关论文
共 15 条
  • [1] Bodik Ras, 2006, UCBEECS2006183
  • [2] The University of Florida Sparse Matrix Collection
    Davis, Timothy A.
    Hu, Yifan
    [J]. ACM TRANSACTIONS ON MATHEMATICAL SOFTWARE, 2011, 38 (01):
  • [3] Hu Weihua, 2020, Advances in Neural Information Processing Systems, V33
  • [4] GE-SpMM: General-Purpose Sparse Matrix-Matrix Multiplication on GPUs for Graph Neural Networks
    Huang, Guyue
    Dai, Guohao
    Wang, Yu
    Yang, Huazhong
    [J]. PROCEEDINGS OF SC20: THE INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS (SC20), 2020,
  • [5] Sparsity: Optimization framework for sparse matrix kernels
    Im, EJ
    Yelick, K
    Vuduc, R
    [J]. INTERNATIONAL JOURNAL OF HIGH PERFORMANCE COMPUTING APPLICATIONS, 2004, 18 (01) : 135 - 158
  • [6] An Extended Compression Format for the Optimization of Sparse Matrix-Vector Multiplication
    Karakasis, Vasileios
    Gkountouvas, Theodoros
    Kourtis, Kornilios
    Goumas, Georgios
    Koziris, Nectarios
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2013, 24 (10) : 1930 - 1940
  • [7] A fast and high quality multilevel scheme for partitioning irregular graphs
    Karypis, G
    Kumar, V
    [J]. SIAM JOURNAL ON SCIENTIFIC COMPUTING, 1998, 20 (01) : 359 - 392
  • [8] CSR5: An Efficient Storage Format for Cross-Platform Sparse Matrix-Vector Multiplication
    Liu, Weifeng
    Vinter, Brian
    [J]. PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON SUPERCOMPUTING (ICS'15), 2015, : 339 - 350
  • [9] Merrill D, 2016, SC '16: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, P678, DOI 10.1109/SC.2016.57
  • [10] Graph neural networks meet with distributed graph partitioners and reconciliations
    Mu, Zongshen
    Tang, Siliang
    Zong, Chang
    Yu, Dianhai
    Zhuang, Yueting
    [J]. NEUROCOMPUTING, 2023, 518 : 408 - 417