Language-based reasoning graph neural network for commonsense question answering

被引:0
作者
Yang, Meng [1 ]
Wang, Yihao [1 ,2 ]
Gu, Yu [1 ,2 ]
机构
[1] Sun Yat Sen Univ, Sch Comp Sci & Engn, Guangzhou 510006, Peoples R China
[2] SYSU, Minist Educ, Key Lab Machine Intelligence & Adv Comp, Guangzhou 510006, Peoples R China
基金
中国国家自然科学基金;
关键词
Commonsense QA; Language-based reasoning; External knowledge;
D O I
10.1016/j.neunet.2024.106816
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Language model (LM) has played an increasingly important role in the common-sense understanding and reasoning in the CSQA task (Common Sense Question Answering). However, due to the amount of model parameters, increasing training data helps little in further improving model performance. Introducing external knowledge through graph neural networks (GNNs) proves positive in boosting performance, but exploiting different knowledge sources and capturing contextual information between text and knowledge inside remains a challenge. In this paper, we propose LBR-GNN, a L anguage-Based R easoning G raph N eural N etwork method to address these problems, by representing the question with each answer and external knowledge using a language model and predicting the reasoning score with a designed language-based GNN. Our LBR-GNN will first regulate external knowledge into a consistent textual form and encode it using a standard LM to capture the contextual information. Then, we build a graph neural network using the encoded information, especially the language-level edge representation. Finally, we design a novel edge aggregation method to select the edge information for GNN update and the language-guided GNN reasoning. We assess the performance of LBRGNN across the CommonsenseQA, CommonsenseQA-IH, and OpenBookQA datasets. Our evaluation reveals a performance boost of more than 5% compared to the state-of-the-art methods on the CSQA dataset, achieved with a similar number of additional parameters.
引用
收藏
页数:13
相关论文
共 57 条
[31]  
Pennington J, 2014, P 2014 C EMPIRICAL M, P1532, DOI [DOI 10.3115/V1/D14-1162, 10.3115/v1/D14-1162]
[32]  
Peters ME, 2018, 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), P1499, DOI 10.5771/9783845286846
[33]   GNN-SubNet: disease subnetwork detection with explainable graph neural networks [J].
Pfeifer, Bastian ;
Saranti, Anna ;
Holzinger, Andreas .
BIOINFORMATICS, 2022, 38 :ii120-ii126
[34]  
Raffel C, 2020, J MACH LEARN RES, V21
[35]  
Rajani NF, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P4932
[36]   Modeling Relational Data with Graph Convolutional Networks [J].
Schlichtkrull, Michael ;
Kipf, Thomas N. ;
Bloem, Peter ;
van den Berg, Rianne ;
Titov, Ivan ;
Welling, Max .
SEMANTIC WEB (ESWC 2018), 2018, 10843 :593-607
[37]   NPI-GNN: Predicting ncRNA-protein interactions with deep graph neural networks [J].
Shen, Zi-Ang ;
Luo, Tao ;
Zhou, Yuan-Ke ;
Yu, Han ;
Du, Pu-Feng .
BRIEFINGS IN BIOINFORMATICS, 2021, 22 (05)
[38]   Point-GNN: Graph Neural Network for 3D Object Detection in a Point Cloud [J].
Shi, Weijing ;
Rajkumar, Ragunathan .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :1708-1716
[39]  
Singh P., 2002, 18 AAAI C ART INT AA
[40]  
Speer R, 2018, Arxiv, DOI arXiv:1612.03975