Pruning graph neural networks by evaluating edge properties

被引:9
作者
Wang, Li [1 ,2 ]
Huang, Wei [2 ]
Zhang, Miao [3 ]
Pan, Shirui [4 ]
Chang, Xiaojun [2 ]
Su, Steven Weidong [1 ,2 ]
机构
[1] Shandong First Med Univ & Shandong Acad Med Sci, Coll Artif Intelligence & Big Data Med Sci, 6699 Qingdao Rd, Jinan 250000, Shandong, Peoples R China
[2] Univ Technol Sydney, Fac Engn & IT, 81 Broadway, Ultimo, NSW 2007, Australia
[3] Aalborg Univ, Comp Sci Dept, Fredrik Bajers Vej 7K, DK-9220 Aalborg, Denmark
[4] Griffith Univ, Sch Informat & Commun Technol, Parklands Dr, Southport, Qld 4222, Australia
关键词
Graph neural networks; Network pruning; Model compression;
D O I
10.1016/j.knosys.2022.109847
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The emergence of larger and deeper graph neural networks (GNNs) makes their training and inference increasingly expensive. Existing GNN pruning methods simultaneously prune the graph adjacency matrix and the model weights on a pretrained neural network by directly leveraging the lottery-ticket hypothesis, but the benefits of such methods are mainly via weight pruning, and methods based on saliency metrics struggle to outperform random pruning when pruning only the graph adjacency matrix. This motivates us to use different scoring standards for graph edges and network weights during GNN pruning. Thus, rather than measuring the importance of graph edges based on saliency metrics, we formulate the performance of GNNs mathematically with respect to the properties of their edges, elucidating how the performance drop can be avoided by pruning negative edges and nonbridges. This leads to our simple but effective two-step method for GNN pruning, leveraging the saliency metrics for the network pruning while sparsifying the graph with preservation of the loss performance. Experimental results show the effectiveness and efficiency of the proposed method on both small-scale graph datasets (Cora, Citeseer, and PubMed) and a large-scale dataset (Ogbn-ArXiv), where our method saves up to 98% of floating-point operations per second (FLOPs) on the small graphs and 94% of FLOPs on the large one, with no significant drop in accuracy. (c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页数:10
相关论文
共 50 条
[41]   Torsion Graph Neural Networks [J].
Shen, Cong ;
Liu, Xiang ;
Luo, Jiawei ;
Xia, Kelin .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (04) :2946-2956
[42]   STOCHASTIC GRAPH NEURAL NETWORKS [J].
Gao, Zhan ;
Isufi, Elvin ;
Ribeiro, Alejandro .
2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, :9080-9084
[43]   Polarized Graph Neural Networks [J].
Fang, Zheng ;
Xu, Lingjun ;
Song, Guojie ;
Long, Qingqing ;
Zhang, Yingxue .
PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, :1404-1413
[44]   Convolutional Graph Neural Networks [J].
Gama, Fernando ;
Marques, Antonio G. ;
Leus, Geert ;
Ribeiro, Alejandro .
CONFERENCE RECORD OF THE 2019 FIFTY-THIRD ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, 2019, :452-456
[45]   ConveXplainer for Graph Neural Networks [J].
Pereira, Tamara A. ;
Nascimento, Erik Jhones F. ;
Mesquita, Diego ;
Souza, Amauri H. .
INTELLIGENT SYSTEMS, PT II, 2022, 13654 :588-600
[46]   Clenshaw Graph Neural Networks [J].
Guo, Yuhe ;
Wei, Zhewei .
PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, :614-625
[47]   AGGREGATION GRAPH NEURAL NETWORKS [J].
Gama, Fernando ;
Marques, Antonio G. ;
Ribeiro, Alejandro ;
Leus, Geert .
2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, :4943-4947
[48]   Stochastic Graph Neural Networks [J].
Gao, Zhan ;
Isufi, Elvin ;
Ribeiro, Alejandro .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2021, 69 :4428-4443
[49]   Survey on Graph Neural Networks [J].
Gkarmpounis, Georgios ;
Vranis, Christos ;
Vretos, Nicholas ;
Daras, Petros .
IEEE ACCESS, 2024, 12 :128816-128832
[50]   Beyond graph neural networks with lifted relational neural networks [J].
Sourek, Gustav ;
Zelezny, Filip ;
Kuzelka, Ondrej .
MACHINE LEARNING, 2021, 110 (07) :1695-1738