An Attention-Based Graph Neural Network for Heterogeneous Structural Learning

被引:0
作者
Hong, Huiting [1 ]
Guo, Hantao [1 ,2 ]
Lin, Yucheng [1 ]
Yang, Xiaoqing [1 ]
Li, Zang [1 ]
Ye, Jieping [1 ]
机构
[1] Didi Chuxing, AI Labs, Beijing, Peoples R China
[2] Peking Univ, Beijing, Peoples R China
来源
THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE | 2020年 / 34卷
关键词
MODEL;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we focus on graph representation learning of heterogeneous information network (HIN), in which various types of vertices are connected by various types of relations. Most of the existing methods conducted on HIN revise homogeneous graph embedding models via meta-paths to learn low-dimensional vector space of HIN. In this paper, we propose a novel Heterogeneous Graph Structural Attention Neural Network (HetSANN) to directly encode structural information of HIN without meta-path and achieve more informative representations. With this method, domain experts will not be needed to design meta-path schemes and the heterogeneous information can be processed automatically by our proposed model. Specifically, we implicitly represent heterogeneous information using the following two methods: 1) we model the transformation between heterogeneous vertices through a projection in low-dimensional entity spaces; 2) afterwards, we apply the graph neural network to aggregate multi-relational information of projected neighborhood by means of attention mechanism. We also present three extensions of HetSANN, i.e.. voices-sharing product attention for the pairwise relationships in HIN, cycle-consistency loss to retain the transformation between heterogeneous entity spaces, and multi-task learning with full use of information. The experiments conducted on three public datasets demonstrate that our proposed models achieve significant and consistent improvements compared to state-of-the-art solutions.
引用
收藏
页码:4132 / 4139
页数:8
相关论文
共 27 条
[1]   A Bayesian information theoretic model of learning to learn via multiple task sampling [J].
Baxter, J .
MACHINE LEARNING, 1997, 28 (01) :7-39
[2]   BACK-TRANSLATION FOR CROSS-CULTURAL RESEARCH [J].
BRISLIN, RW .
JOURNAL OF CROSS-CULTURAL PSYCHOLOGY, 1970, 1 (03) :185-216
[3]   Task-Guided and Path-Augmented Heterogeneous Network Embedding for Author Identification [J].
Chen, Ting ;
Sun, Yizhou .
WSDM'17: PROCEEDINGS OF THE TENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2017, :295-304
[4]   metapath2vec: Scalable Representation Learning for Heterogeneous Networks [J].
Dong, Yuxiao ;
Chawla, Nitesh V. ;
Swami, Ananthram .
KDD'17: PROCEEDINGS OF THE 23RD ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2017, :135-144
[5]  
Gori M, 2005, IEEE IJCNN, P729
[6]  
Hamilton WL, 2017, ADV NEUR IN, V30
[7]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[8]  
Ji M, 2010, LECT NOTES ARTIF INT, V6321, P570
[9]  
Kingma DP., 2017, A method for stochastic optimization, DOI DOI 10.48550/ARXIV.1412.6980
[10]  
Kipf Thomas N., 2017, P 5 INT C LEARNING R, P1, DOI DOI 10.48550/ARXIV.1609.02907