Evaluating explainability for graph neural networks

被引:0
|
作者
Chirag Agarwal
Owen Queen
Himabindu Lakkaraju
Marinka Zitnik
机构
[1] Media and Data Science Research Lab,Department of Biomedical Informatics
[2] Adobe,Department of Electrical Engineering and Computer Science
[3] Harvard University,Department of Computer Science
[4] University of Tennessee,undefined
[5] Harvard Business School,undefined
[6] Harvard Data Science Initiative,undefined
[7] Harvard University,undefined
[8] Broad Institute of MIT and Harvard,undefined
来源
Scientific Data | / 10卷
关键词
D O I
暂无
中图分类号
学科分类号
摘要
As explanations are increasingly used to understand the behavior of graph neural networks (GNNs), evaluating the quality and reliability of GNN explanations is crucial. However, assessing the quality of GNN explanations is challenging as existing graph datasets have no or unreliable ground-truth explanations. Here, we introduce a synthetic graph data generator, ShapeGGen, which can generate a variety of benchmark datasets (e.g., varying graph sizes, degree distributions, homophilic vs. heterophilic graphs) accompanied by ground-truth explanations. The flexibility to generate diverse synthetic datasets and corresponding ground-truth explanations allows ShapeGGen to mimic the data in various real-world areas. We include ShapeGGen and several real-world graph datasets in a graph explainability library, GraphXAI. In addition to synthetic and real-world graph datasets with ground-truth explanations, GraphXAI provides data loaders, data processing functions, visualizers, GNN model implementations, and evaluation metrics to benchmark GNN explainability methods.
引用
收藏
相关论文
共 50 条
  • [41] Graphs, Convolutions, and Neural Networks: From Graph Filters to Graph Neural Networks
    Gama, Fernando
    Isufi, Elvin
    Leus, Geert
    Ribeiro, Alejandro
    IEEE SIGNAL PROCESSING MAGAZINE, 2020, 37 (06) : 128 - 138
  • [42] Neural Pooling for Graph Neural Networks
    Harsha, Sai Sree
    Mishra, Deepak
    PATTERN RECOGNITION AND MACHINE INTELLIGENCE, PREMI 2021, 2024, 13102 : 171 - 180
  • [43] Graphon Neural Networks and the Transferability of Graph Neural Networks
    Ruiz, Luana
    Chamon, Luiz F. O.
    Ribeiro, Alejandro
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [44] Explainability of Neural Networks for Symbol Detection in Molecular Communication Channels
    Gomez, Jorge Torres
    Hofmann, Pit
    Fitzek, Frank H. P.
    Dressler, Falko
    IEEE TRANSACTIONS ON MOLECULAR BIOLOGICAL AND MULTI-SCALE COMMUNICATIONS, 2023, 9 (03): : 323 - 328
  • [45] A concept-aware explainability method for convolutional neural networks
    Gurkan, Mustafa Kagan
    Arica, Nafiz
    Vural, Fatos T. Yarman
    MACHINE VISION AND APPLICATIONS, 2025, 36 (02)
  • [46] Improving explainability results of convolutional neural networks in microscopy images
    Kallipolitis, Athanasios
    Yfantis, Panayiotis
    Maglogiannis, Ilias
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (29): : 21535 - 21553
  • [47] Explainability of deep neural networks for MRI analysis of brain tumors
    Zeineldin, Ramy A.
    Karar, Mohamed E.
    Elshaer, Ziad
    Coburger, Jan
    Wirtz, Christian R.
    Burgert, Oliver
    Mathis-Ullrich, Franziska
    INTERNATIONAL JOURNAL OF COMPUTER ASSISTED RADIOLOGY AND SURGERY, 2022, 17 (09) : 1673 - 1683
  • [48] Explainability of deep neural networks for MRI analysis of brain tumors
    Ramy A. Zeineldin
    Mohamed E. Karar
    Ziad Elshaer
    ·Jan Coburger
    Christian R. Wirtz
    Oliver Burgert
    Franziska Mathis-Ullrich
    International Journal of Computer Assisted Radiology and Surgery, 2022, 17 : 1673 - 1683
  • [49] Some Shades of Grey! - Interpretability and Explainability of Deep Neural Networks
    Dengel, Andreas
    PROCEEDINGS OF THE ACM WORKSHOP ON CROSSMODAL LEARNING AND APPLICATION (WCRML'19), 2019, : 1 - 1
  • [50] Improving explainability results of convolutional neural networks in microscopy images
    Athanasios Kallipolitis
    Panayiotis Yfantis
    Ilias Maglogiannis
    Neural Computing and Applications, 2023, 35 : 21535 - 21553