Improving performance and efficiency of Graph Neural Networks by injective aggregation

被引:38
作者
Dong, Wei [1 ]
Wu, Junsheng [2 ]
Zhang, Xinwan [2 ]
Bai, Zongwen [3 ]
Wang, Peng [4 ]
Wozniak, Marcin [5 ]
机构
[1] Northwestern Polytech Univ, Sch Comp Sci & Engn, Xian, Peoples R China
[2] Northwestern Polytech Univ, Sch Software, Xian, Peoples R China
[3] Yanan Univ, Sch Phys & Elect Informat, Shaanxi Key Lab Intelligent Proc Big Energy Data, Yanan, Peoples R China
[4] Univ Wollongong, Sch Comp & Informat Technol, Wollongong, NSW 2522, Australia
[5] Silesian Tech Univ, Fac Appl Math, Kaszubsa 23, PL-44100 Gliwice, Poland
基金
中国国家自然科学基金;
关键词
Graph Neural Networks; Aggregation function; Aggregation matrix; Injectivity; Traffic state prediction;
D O I
10.1016/j.knosys.2022.109616
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Aggregation functions are regarded as the multiplication between an aggregation matrix and node embeddings, based on which a full rank matrix can enhance representation capacity of Graph Neural Networks (GNNs). In this work, we fill this research gap based on the full rank aggregation matrix and its functional form, i.e., the injective aggregation function, and state that injectivity is necessary to guarantee the rich representation capacity to GNNs. To this end, we conduct theoretical injectivity analysis for the typical feature aggregation methods and provide inspiring solutions on turning the non-injective aggregation functions into injective versions. Based on our injective aggregation functions, we create various GNN structures by combining the aggregation functions with the other ingredient of GNNs, node feature encoding, in different ways. Following these structures, we highlight that by using our injective aggregation function entirely as a pre-processing step before applying independent node feature learning, we can simultaneously achieve satisfactory performance and computational efficiency on the large-scale graph-based traffic data for traffic state prediction tasks. Through comprehensive experiments on standard node classification benchmarks and practical traffic state data (for Chengdu and Xi'an cities), we show that the representation capacity of GNNs can be improved by using our injective aggregation functions just by changing the model in simple operations. (C) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页数:15
相关论文
共 43 条
[1]  
Amodei D, 2016, PR MACH LEARN RES, V48
[2]  
Babai L., 1979, 20th Annual Symposium of Foundations of Computer Science, P39, DOI 10.1109/SFCS.1979.8
[3]  
Bresson X, 2018, Arxiv, DOI arXiv:1711.07553
[4]  
Cao Yue, 2019, Adv Neural Inf Process Syst, V32, P15044
[5]  
Chen J., 2018, INT C LEARNING REPRE
[6]  
Chen JF, 2018, PR MACH LEARN RES, V80
[7]   Multi-Label Image Recognition with Graph Convolutional Networks [J].
Chen, Zhao-Min ;
Wei, Xiu-Shen ;
Wang, Peng ;
Guo, Yanwen .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :5172-5181
[8]   Cluster-GCN: An Efficient Algorithm for Training Deep and Large Graph Convolutional Networks [J].
Chiang, Wei-Lin ;
Liu, Xuanqing ;
Si, Si ;
Li, Yang ;
Bengio, Samy ;
Hsieh, Cho-Jui .
KDD'19: PROCEEDINGS OF THE 25TH ACM SIGKDD INTERNATIONAL CONFERENCCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2019, :257-266
[9]  
Defferrard M, 2016, ADV NEUR IN, V29
[10]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171