Addressing imbalance in graph datasets: Introducing GATE-GNN with graph ensemble weight attention and transfer learning for enhanced node classification

被引:8
作者
Fofanah, Abdul Joseph [1 ]
Chen, David [1 ]
Wen, Lian [1 ]
Zhang, Shaoyang [2 ]
机构
[1] Griffith Univ, Sch Informat & Commun Technol, 170 Kessels Rd, Brisbane, Qld 4111, Australia
[2] Changan Univ, Sch Informat & Commun Technol, Xian, Peoples R China
关键词
Classification; Ensemble method; Ensemble attention; Graph neural network; Transfer learning method; Imbalanced dataset; SMOTE;
D O I
10.1016/j.eswa.2024.124602
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Significant challenges arise when Graph Neural Networks (GNNs) try to deal with uneven data. Specifically in signed and weighted graph structures. This makes classification tasks less effective. Within the GNN context, researchers have found traditional solutions like resampling, reweighting, and synthetic sample generation to be inadequate. GATE-GNN, a novel architecture designed specifically for imbalanced datasets, overcomes these limitations. GATE-GNN integrates an ensemble of network modules that harness the spatial features of graph networks and effectively utilise embedding information from earlier layers. This unique approach not only bolsters generalisation by reducing volatility. It also refines the optimisation algorithm, resulting in more accurate and stable classification outcomes. We rigorously tested the effectiveness of GATE-GNN on four widely recognised datasets: Cora, NELL, Citeseer, and PubMed. We performed a comparative analysis against established methods such as Graph Convolutional Networks (GCN), Graph Sample and Aggregate (GraphSAGE), Propagation Multilayer Perceptron PMLP), Imbalanced Node Sampling GNN (INS-GNN), GNN-Curriculum Learning (GNN-CL) and Graph Attention Networks (GAT). Empirical results demonstrate that GATE-GNN significantly outperforms these existing models, achieving an average improvement in classification accuracy of approximately 5%-10% over the previous best results. Additionally, GATE-GNN presents a marked reduction in training time. This underscores its efficiency and suitability for practical applications in imbalanced graph data scenarios. Implementation of the proposed GATE-GNN can be accessed here https://github.com/afofanah/ GATE-GNN.
引用
收藏
页数:16
相关论文
共 44 条
[1]  
Abu-El-Haija S, 2019, PR MACH LEARN RES, V115, P841
[2]   Feature Selection and Extraction for Graph Neural Networks [J].
Acharya, Deepak Bhaskar ;
Zhang, Huaming .
ACMSE 2020: PROCEEDINGS OF THE 2020 ACM SOUTHEAST CONFERENCE, 2020, :252-255
[3]   Attribute-based Decision Graphs: A framework for multiclass data classification [J].
Bertini Junior, Joao Roberto ;
Nicoletti, Maria do Carmo ;
Zhao, Liang .
NEURAL NETWORKS, 2017, 85 :69-84
[4]   Class-Balanced Loss Based on Effective Number of Samples [J].
Cui, Yin ;
Jia, Menglin ;
Lin, Tsung-Yi ;
Song, Yang ;
Belongie, Serge .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9260-9269
[5]   MFDGCN: Multi-Stage Spatio-Temporal Fusion Diffusion Graph Convolutional Network for Traffic Prediction [J].
Cui, Zhengyan ;
Zhang, Junjun ;
Noh, Giseop ;
Park, Hyun Jun .
APPLIED SCIENCES-BASEL, 2022, 12 (05)
[6]   Geometric SMOTE a geometrically enhanced drop-in replacement for SMOTE [J].
Douzas, Georgios ;
Bacao, Fernando .
INFORMATION SCIENCES, 2019, 501 :118-135
[7]   Handling Imbalanced Data in Road Crash Severity Prediction by Machine Learning Algorithms [J].
Fiorentini, Nicholas ;
Losa, Massimo .
INFRASTRUCTURES, 2020, 5 (07)
[8]   Bilinear Graph Networks for Visual Question Answering [J].
Guo, Dalu ;
Xu, Chang ;
Tao, Dacheng .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (02) :1023-1034
[9]  
Guo ZC, 2023, Arxiv, DOI arXiv:2210.05920
[10]  
Hamilton WL, 2017, ADV NEUR IN, V30