Graph neural networks meet with distributed graph partitioners and reconciliations

被引:3
作者
Mu, Zongshen [1 ]
Tang, Siliang [1 ]
Zong, Chang [1 ]
Yu, Dianhai [2 ]
Zhuang, Yueting [1 ]
机构
[1] Zhejiang Univ, Coll Comp Sci, DCD Lab, Hangzhou 310013, Peoples R China
[2] Baidu Inc, Beijing 100094, Peoples R China
关键词
Distributed GNNs; Graph Partitioning; Knowledge Distillation; Graph Contrastive Learning;
D O I
10.1016/j.neucom.2022.09.096
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph neural networks (GNNs) have shown great success in various applications. As real-world graphs are large, training GNNs in distributed systems is desirable. In current training schemes, their edge par-titioning strategies have a strong impact on the performance of GNNs for the unbalanced influence of high-degree nodes and the damaged neighbor integrity of low-degree nodes. Meanwhile, a lack of recon-ciliations of different local models leads to converging up and down across workers. In this work, we design DEPR, a suitable framework for distributed GNN training. We propose a degree-sensitive edge par-titioning with influence-balancing and locality-preserving to adapt distributed GNNs training by follow-ing an owner-compute rule (each partition performs all the computations involving data that it owns). And then knowledge distillation and contrastive learning are used to reconcile the fusion of local models and boost convergence. We show in extensive empirical experiments on the node classification task of three large-scale graph datasets (Reddit, Amazon, and OGB-Products) that DEPR achieves 2x speedup of convergence and get absolute up 3.97 performance improvement of F1-micro score compared to DistDGL. (c) 2022 Published by Elsevier B.V.
引用
收藏
页码:408 / 417
页数:10
相关论文
共 39 条
[1]  
Agarwal A, 2014, J MACH LEARN RES, V15, P1111
[2]  
Anil R, 2020, Arxiv, DOI arXiv:1804.03235
[3]   Parallel mesh partitioning based on space filling curves [J].
Borrell, R. ;
Cajas, J. C. ;
Mira, D. ;
Taha, A. ;
Koric, S. ;
Vazquez, M. ;
Houzeaux, G. .
COMPUTERS & FLUIDS, 2018, 173 :264-272
[4]  
Chen J, 2018, Arxiv, DOI arXiv:1801.10247
[5]  
Chen T, 2020, PR MACH LEARN RES, V119
[6]   Cluster-GCN: An Efficient Algorithm for Training Deep and Large Graph Convolutional Networks [J].
Chiang, Wei-Lin ;
Liu, Xuanqing ;
Si, Si ;
Li, Yang ;
Bengio, Samy ;
Hsieh, Cho-Jui .
KDD'19: PROCEEDINGS OF THE 25TH ACM SIGKDD INTERNATIONAL CONFERENCCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2019, :257-266
[7]  
Cui P., IEEE Transactions on Knowledge and Data Engineering
[8]   Diversity with Cooperation: Ensemble Methods for Few-Shot Classification [J].
Dvornik, Nikita ;
Schmid, Cordelia ;
Mairal, Julien .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :3722-3730
[9]   Multi-scale features based interpersonal relation recognition using higher-order graph neural network [J].
Gao, Jianjun ;
Qing, Linbo ;
Li, Lindong ;
Cheng, Yongqiang ;
Peng, Yonghong .
NEUROCOMPUTING, 2021, 456 :243-252
[10]  
github.com, 2021, PADDLE GRAPH LEARNIN