QUINT: Node Embedding Using Network Hashing

被引:4
作者
Bera, Debajyoti [1 ]
Pratap, Rameshwar [2 ]
Verma, Bhisham Dev [2 ,3 ]
Sen, Biswadeep
Chakraborty, Tanmoy [1 ]
机构
[1] IIT Delhi, New Delhi 110020, Delhi, India
[2] Indian Inst Technol, Mandi 175005, Himachal Prades, India
[3] Natl Univ Singapore, Dept Comp Sci, Singapore 119077, Singapore
关键词
Task analysis; Training; Sparse matrices; Optimization; Linear matrix inequalities; Statistical analysis; Standards; Network embedding; node classification; link prediction; sparse network; binary sketch; dimensionality reduction;
D O I
10.1109/TKDE.2021.3111997
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Representation learning using network embedding has received tremendous attention due to its efficacy to solve downstream tasks. Popular embedding methods (such as deepwalk, node2vec, LINE) are based on a neural architecture, thus unable to scale on large networks both in terms of time and space usage. Recently, we proposed BinSketch, a sketching technique for compressing binary vectors to binary vectors. In this paper, we show how to extend BinSketch and use it for network hashing. Our proposal named QUINT is built upon BinSketch, and it embeds nodes of a sparse network onto a low-dimensional space using simple bit-wise operations. QUINT is the first of its kind that provides tremendous gain in terms of speed and space usage without compromising much on the accuracy of the downstream tasks. Extensive experiments are conducted to compare QUINT with seven state-of-the-art network embedding methods for two end tasks - link prediction and node classification. We observe huge performance gain for QUINT in terms of speedup (up to 7000x) and space saving (up to 800x) due to its bit-wise nature to obtain node embedding. Moreover, QUINT is a consistent top-performer for both the tasks among the baselines across all the datasets. Our empirical observations are backed by rigorous theoretical analysis to justify the effectiveness of QUINT. In particular, we prove that QUINT retains enough structural information which can be used further to approximate many topological properties of networks with high confidence.
引用
收藏
页码:2987 / 3000
页数:14
相关论文
共 51 条
  • [1] [Anonymous], 1984, C MOD AN PROB 1984
  • [2] [Anonymous], 2015, P 24 ACM INT C INF K
  • [3] [Anonymous], 2021, About us
  • [4] [Anonymous], 2009, P 18 ACM C INF KNOWL
  • [5] [Anonymous], 2008, 21 INT C NEURAL INF
  • [6] A Comprehensive Survey of Graph Embedding: Problems, Techniques, and Applications
    Cai, HongYun
    Zheng, Vincent W.
    Chang, Kevin Chen-Chuan
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2018, 30 (09) : 1616 - 1637
  • [7] Fast and Accurate Network Embeddings via Very Sparse Random Projection
    Chen, Haochen
    Sultan, Syed Fahad
    Tian, Yingtao
    Chen, Muhao
    Skiena, Steven
    [J]. PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, : 399 - 408
  • [8] SSNE: Effective Node Representation for Link Prediction in Sparse Networks
    Chen, Ming-Ren
    Huang, Ping
    Lin, Yu
    Cai, Shi-Min
    [J]. IEEE ACCESS, 2021, 9 : 57874 - 57885
  • [9] Fast linear algebra is stable
    Demmel, James
    Dumitriu, Ioana
    Holtz, Olga
    [J]. NUMERISCHE MATHEMATIK, 2007, 108 (01) : 59 - 91
  • [10] metapath2vec: Scalable Representation Learning for Heterogeneous Networks
    Dong, Yuxiao
    Chawla, Nitesh V.
    Swami, Ananthram
    [J]. KDD'17: PROCEEDINGS OF THE 23RD ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2017, : 135 - 144