A Simple Yet Effective Layered Loss for Pre-Training of Network Embedding

被引:6
作者
Chen, Junyang [1 ]
Li, Xueliang [1 ]
Li, Yuanman [2 ]
Li, Paul [3 ]
Wang, Mengzhu [4 ]
Zhang, Xiang [4 ]
Gong, Zhiguo [5 ]
Wu, Kaishun [1 ]
Leung, Victor C. M. [1 ]
机构
[1] Shenzhen Univ, Coll Comp Sci & Software Engn, Shenzhen 518060, Peoples R China
[2] Shenzhen Univ, Coll Elect & Informat Engn, Guangdong Key Lab Intelligent Informat Proc, Shenzhen 518060, Peoples R China
[3] Baidu Res, Sunnyvale, CA 94089 USA
[4] Natl Univ Def Technol, Coll Comp, Changsha 564211, Peoples R China
[5] Univ Macau, Dept Comp Informat Sci, State Key Lab Internet Things Smart City, Macau 999078, Peoples R China
来源
IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING | 2022年 / 9卷 / 03期
基金
中国国家自然科学基金;
关键词
Task analysis; Graph neural networks; Training; Propagation losses; Message passing; Aggregates; Web and internet services; layered loss; network embedding; pre-training of unlabeled nodes;
D O I
10.1109/TNSE.2022.3153643
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Pre-training of network embedding aims to encode unlabeled node proximity into a low-dimensional space, where nodes are close to their neighbors while being far from negative samples. In recent years, Graph Neural Networks have shown groundbreaking performance in semi-supervised learning on the node classification and link prediction tasks. However, because of their inherent information aggregation pattern, almost all these methods can only obtain inferior embedding results in the pre-training of the unlabeled nodes. The margins between a target node and its multi-hop neighbors become hard distinguishable during node message aggregation. To address this problem, we propose a simple yet effective layered loss to combine with a graph attention network, dubbed as LlossNet, for pre-training. We regard the proximity of a target node and its two-hop neighbors as a unit (called a unit graph), where a target node is needed to be more closer to its direct neighbor than its two-hop neighbors. As such, LlossNet would be able to preserve the margins of nodes in the learned embedding space. Experimental results of various downstream tasks including classification and clustering demonstrate the effectiveness of our method on learning discriminative node representations.
引用
收藏
页码:1827 / 1837
页数:11
相关论文
共 35 条
[11]   node2vec: Scalable Feature Learning for Networks [J].
Grover, Aditya ;
Leskovec, Jure .
KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, :855-864
[12]  
Hamilton WL, 2017, ADV NEUR IN, V30
[13]  
Han J, 1995, LECT NOTES COMPUT SC, V930, P195
[14]   Deep Embedding Network for Clustering [J].
Huang, Peihao ;
Huang, Yan ;
Wang, Wei ;
Wang, Liang .
2014 22ND INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2014, :1532-1537
[15]  
Kingma DP, 2014, ADV NEUR IN, V27
[16]  
Kipf T.N., 2017, INT C LEARN REPR ICL
[17]   Gradient-based learning applied to document recognition [J].
Lecun, Y ;
Bottou, L ;
Bengio, Y ;
Haffner, P .
PROCEEDINGS OF THE IEEE, 1998, 86 (11) :2278-2324
[18]   Reducing the Sampling Complexity of Topic Models [J].
Li, Aaron Q. ;
Ahmed, Amr ;
Ravi, Sujith ;
Smola, Alexander J. .
PROCEEDINGS OF THE 20TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING (KDD'14), 2014, :891-900
[19]   Automating the construction of internet portals with machine learning [J].
McCallum, AK ;
Nigam, K ;
Rennie, J ;
Seymore, K .
INFORMATION RETRIEVAL, 2000, 3 (02) :127-163
[20]  
Mikolov T., 2013, P WORKSH ICLR