DistMILE: A Distributed Multi-Level Framework for Scalable Graph Embedding

被引:8
作者
He, Yuntian [1 ]
Gurukar, Saket [1 ]
Kousha, Pouya [1 ]
Subramoni, Hari [1 ]
Panda, Dhabaleswar K. [1 ]
Parthasarathy, Srinivasan [1 ]
机构
[1] Ohio State Univ, Dept Comp Sci & Engn, Columbus, OH 43210 USA
来源
2021 IEEE 28TH INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPUTING, DATA, AND ANALYTICS (HIPC 2021) | 2021年
基金
美国国家科学基金会;
关键词
Graph Embedding; High-Performance Computing; Distributed Machine Learning; Multi-Level Framework; WAY PARTITIONING SCHEME;
D O I
10.1109/HiPC53243.2021.00042
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Scalable graph embedding on large networks is challenging because of the complexity of graph structures and limited computing resources. Recent research shows that the multi-level framework can enhance the scalability of graph embedding methods with little loss of quality. In general, methods using this framework first coarsen the original graph into a series of smaller graphs then learn the representations of the original graph from them in an efficient manner. However, to the best of our knowledge, most multi-level based methods do not have a parallel implementation. Meanwhile, the emergence of high-performance computing for machine learning provides an opportunity to boast graph embedding by distributed computing. In this paper, we propose a Distributed Multi-Level Embedding (DistMILE(1)) framework to further improve the scalability of graph embedding. DistMILE leverages a novel shared-memory parallel algorithm for graph coarsening and a distributed training paradigm for embedding refinement. With the advantage of high-performance computing techniques, DistMILE can smoothly scale different base embedding methods over large networks. Our experiments demonstrate that DistMILE learns representations of similar quality with respect to other baselines, while reduces the time of learning embeddings on large-scale networks to hours. Results show that DistMILE can achieve up to 28 x speedup compared with a popular multi-level embedding framework MILE and expedite existing embedding methods with 10x speedup.
引用
收藏
页码:282 / 291
页数:10
相关论文
共 34 条
[1]  
Akyildiz T.A., 2020, PROC INT C PARALLEL, P1
[2]  
[Anonymous], 2020, Journal of Computational Science, page, P101208
[3]  
[Anonymous], 2016, TensorFlow: large-scale machine learning on heterogeneous distributed systems
[4]  
[Anonymous], 2002, P 34 ANN ACM S THEOR, DOI DOI 10.1145/509907.509965
[5]  
Chen HC, 2018, AAAI CONF ARTIF INTE, P2127
[6]   DIANNE: a modular framework for designing, training and deploying deep neural networks on heterogeneous distributed infrastructure [J].
De Coninck, Elias ;
Bohez, Steven ;
Leroux, Sam ;
Verbelen, Tim ;
Vankeirsbilck, Bert ;
Simoens, Pieter ;
Dhoedt, Bart .
JOURNAL OF SYSTEMS AND SOFTWARE, 2018, 141 :52-65
[7]  
Dean J., 2012, ADV NEURAL INFORM PR
[8]  
Deng C., 2020, P INT C LEARN REPR
[9]  
Gibiansky A, 2017, BRINGING HPC TECHNIQ
[10]  
Goyal P., 2017, Technical Report