Sparse Graph Attention Networks

被引:80
|
作者
Ye, Yang [1 ]
Ji, Shihao [1 ]
机构
[1] Georgia State Univ, Dept Comp Sci, Atlanta, GA 30303 USA
关键词
Graph neural networks; attention networks; sparsity learning;
D O I
10.1109/TKDE.2021.3072345
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph Neural Networks (GNNs) have proved to be an effective representation learning framework for graph-structured data, and have achieved state-of-the-art performance on many practical predictive tasks, such as node classification, link prediction and graph classification. Among the variants of GNNs, Graph Attention Networks (GATs) learn to assign dense attention coefficients over all neighbors of a node for feature aggregation, and improve the performance of many graph learning tasks. However, real-world graphs are often very large and noisy, and GATs are prone to overfitting if not regularized properly. Even worse, the local aggregation mechanism of GATs may fail on disassortative graphs, where nodes within local neighborhood provide more noise than useful information for feature aggregation. In this paper, we propose Sparse Graph Attention Networks (SGATs) that learn sparse attention coefficients under an $L_0$L0-norm regularization, and the learned sparse attentions are then used for all GNN layers, resulting in an edge-sparsified graph. By doing so, we can identify noisy/task-irrelevant edges, and thus perform feature aggregation on most informative neighbors. Extensive experiments on synthetic and real-world (assortative and disassortative) graph learning benchmarks demonstrate the superior performance of SGATs. In particular, SGATs can remove about 50-80 percent edges from large assortative graphs, such as PPI and Reddit, while retaining similar classification accuracies. On disassortative graphs, SGATs prune majority of noisy edges and outperform GATs in classification accuracies by significant margins. Furthermore, the removed edges can be interpreted intuitively and quantitatively. To the best of our knowledge, this is the first graph learning algorithm that shows significant redundancies in graphs and edge-sparsified graphs can achieve similar (on assortative graphs) or sometimes higher (on disassortative graphs) predictive performances than original graphs. Our code is available at https://github.com/Yangyeeee/SGAT.
引用
收藏
页码:905 / 916
页数:12
相关论文
共 50 条
  • [1] Optimization and Interpretability of Graph Attention Networks for Small Sparse Graph Structures in Automotive Applications
    Neumeier, Marion
    Tollkuehn, Andreas
    Dorn, Sebastian
    Botsch, Michael
    Utschick, Wolfgang
    2023 IEEE INTELLIGENT VEHICLES SYMPOSIUM, IV, 2023,
  • [2] Construction safety predictions with multi-head attention graph and sparse accident networks
    Mostofi, Fatemeh
    Togan, Vedat
    AUTOMATION IN CONSTRUCTION, 2023, 156
  • [3] Multistage Superpixel-Guided Hyperspectral Image Classification With Sparse Graph Attention Networks
    Li, Weiming
    Liu, Qikang
    Fan, Shuaishuai
    Bai, Hongyang
    Xin, Mingrui
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [4] Graph Ordering Attention Networks
    Chatzianastasis, Michail
    Lutzeyer, Johannes
    Dasoulas, George
    Vazirgiannis, Michalis
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 6, 2023, : 7006 - 7014
  • [5] A REGULARIZED ATTENTION MECHANISM FOR GRAPH ATTENTION NETWORKS
    Shanthamallu, Uday Shankar
    Jayaraman, J. Thiagarajan
    Spanias, Andreas
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3372 - 3376
  • [6] Graph Oriented Attention Networks
    Amine, Ouardi
    Mestari, Mohammed
    IEEE ACCESS, 2024, 12 : 47057 - 47067
  • [7] Signed Graph Attention Networks
    Huang, Junjie
    Shen, Huawei
    Hou, Liang
    Cheng, Xueqi
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: WORKSHOP AND SPECIAL SESSIONS, 2019, 11731 : 566 - 577
  • [8] Spatial-Temporal Data Inference With Graph Attention Neural Networks in Sparse Mobile Crowdsensing
    Yang, Guisong
    Wen, Panpan
    Liu, Yutong
    Kong, Linghe
    Liu, Yunhuai
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2024, 11 (05): : 4617 - 4626
  • [9] Bi-channel Multiple Sparse Graph Attention Networks for Session-based Recommendation
    Qiao, Shutong
    Zhou, Wei
    Wen, Junhao
    Zhang, Hongyu
    Gao, Min
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 2075 - 2084
  • [10] A unified deep sparse graph attention network for scene graph generation
    Zhou, Hao
    Yang, Yazhou
    Luo, Tingjin
    Zhang, Jun
    Li, Shuohao
    PATTERN RECOGNITION, 2022, 123