Bipartite Graph Coarsening for Text Classification Using Graph Neural Networks

被引:1
作者
dos Santos, Nicolas Roque [1 ]
Minatel, Diego [1 ]
Baria Valejo, Alan Demetrius [2 ]
Lopes, Alneu de A. [1 ]
机构
[1] Univ Sao Paulo, Inst Math & Comp Sci, Sao Carlos, Brazil
[2] Univ Fed Sao Carlos, Dept Comp, Sao Carlos, Brazil
来源
PROGRESS IN PATTERN RECOGNITION, IMAGE ANALYSIS, COMPUTER VISION, AND APPLICATIONS, CIARP 2023, PT I | 2024年 / 14469卷
基金
巴西圣保罗研究基金会;
关键词
Coarsening; Multilevel Optimization; Graph Neural Network; Text Mining;
D O I
10.1007/978-3-031-49018-7_42
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text classification is a fundamental task in Text Mining (TM) with applications ranging from spam detection to sentiment analysis. One of the current approaches to this task is Graph Neural Network (GNN), primarily used to deal with complex and unstructured data. However, the scalability of GNNs is a significant challenge when dealing with large-scale graphs. Multilevel optimization is prominent among the methods proposed to tackle the issues that arise in such a scenario. This approach uses a hierarchical coarsening technique to reduce a graph, then applies a target algorithm to the coarsest graph and projects the output back to the original graph. Here, we propose a novel approach for text classification using GNN. We build a bipartite graph from the input corpus and then apply the coarsening technique of the multilevel optimization to generate ten contracted graphs to analyze the GNN's performance, training time, and memory consumption as the graph is gradually reduced. Although we conducted experiments on text classification, we emphasize that the proposed method is not bound to a specific task and, thus, can be generalized to different problems modeled as bipartite graphs. Experiments on datasets from various domains and sizes show that our approach reduces memory consumption and training time without significantly losing performance.
引用
收藏
页码:589 / 604
页数:16
相关论文
共 32 条
[1]  
[Anonymous], 2012, P 50 ANN M ASS COMP
[2]  
[Anonymous], 2013, P 2013 C EMP METH NA
[3]  
[Anonymous], 2005, P ACL
[4]  
Bruna J., 2014, 2 INT C LEARN REPR I
[5]  
Cai C., 2021, ICLR
[6]   Cluster-GCN: An Efficient Algorithm for Training Deep and Large Graph Convolutional Networks [J].
Chiang, Wei-Lin ;
Liu, Xuanqing ;
Si, Si ;
Li, Yang ;
Bengio, Samy ;
Hsieh, Cho-Jui .
KDD'19: PROCEEDINGS OF THE 25TH ACM SIGKDD INTERNATIONAL CONFERENCCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2019, :257-266
[7]  
Craven M, 1998, FIFTEENTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-98) AND TENTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICAL INTELLIGENCE (IAAI-98) - PROCEEDINGS, P509
[8]  
Ding KZ, 2020, PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), P4927
[9]   Learning Graph Pooling and Hybrid Convolutional Operations for Text Representations [J].
Gao, Hongyang ;
Chen, Yongjun ;
Ji, Shuiwang .
WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, :2743-2749
[10]  
Gilmer J, 2017, PR MACH LEARN RES, V70