Learning graph structure via graph convolutional networks

被引:46
作者
Zhang, Qi [1 ,2 ]
Chang, Jianlong [1 ,2 ]
Meng, Gaofeng [1 ]
Xu, Shibiao [1 ]
Xiang, Shiming [1 ,2 ]
Pan, Chunhong [1 ]
机构
[1] Chinese Acad Sci, Inst Automat, Natl Lab Pattern Recognit, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 100049, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep learning; Graph convolutional neural networks; Graph structure learning; Changeable kernel sizes; NEURAL-NETWORK; SEQUENCE;
D O I
10.1016/j.patcog.2019.06.012
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph convolutional neural networks have aroused more and more attentions on account of the ability to handle the graph-structured data defined on irregular or non-Euclidean domains. Different from the data defined on regular grids, each node in the graph-structured data has different number of neighbors, and the interactions and correlations between nodes vary at different locations, resulting in complex graph structure. However, the existing graph convolutional neural networks generally pay little attention to exploiting the graph structure information. Moreover, most existing graph convolutional neural networks employ the weight sharing strategy which lies on the statistical assumption of stationarity. This assumption is not always verified on the graph-structured data. To address these issues, we propose a method that learns Graph Structure via graph Convolutional Networks (GSCN), which introduces the graph structure parameters measuring the correlation degrees of adjacent nodes. The graph structure parameters are constantly modified the graph structure during the training phase and will help the filters of the proposed method to focus on the relevant nodes in each neighborhood. Meanwhile by combining the graph structure parameters and kernel weights, our method, which relaxes the restriction of weight sharing, is better to handle the graph-structured data of non-stationarity. In addition, the non-linear activation function ReLU and the sparse constraint are employed on the graph structure parameters to promote GSCN to focus on the important links and filter out the insignificant links in each neighborhood. Experiments on various tasks, including text categorization, molecular activity detection, traffic forecasting and skeleton-based action recognition, illustrate the validity of our method. (C) 2019 Elsevier Ltd. All rights reserved.
引用
收藏
页码:308 / 318
页数:11
相关论文
共 35 条
  • [1] [Anonymous], 1989, P ADV NEURAL INFORM
  • [2] [Anonymous], PROC CVPR IEEE
  • [3] [Anonymous], 2015, Nature, DOI [10.1038/nature14539, DOI 10.1038/NATURE14539]
  • [4] [Anonymous], 2017, COMMUN ACM, DOI DOI 10.1145/3065386
  • [5] [Anonymous], 2013, CORR
  • [6] [Anonymous], 2015, P 3 INT C LEARN REPR
  • [7] [Anonymous], 2015, CORR
  • [8] [Anonymous], P INT C MACH LEARN
  • [9] [Anonymous], IEEE T INTELL TRANSP
  • [10] [Anonymous], 2017, P INT C LEARN REPR