Multi-hop Relational Graph Attention Network for Text-to-SQL Parsing

被引:0
作者
Liu, Hu [1 ]
Shi, Yuliang [1 ,2 ]
Zhang, Jianlin [1 ]
Wang, Xinjun [1 ,2 ]
Li, Hui [1 ]
Kong, Fanyu [1 ]
机构
[1] Shandong Univ, Sch Software, Jinan, Peoples R China
[2] Dareway Software Co Ltd, Jinan, Peoples R China
来源
2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN | 2023年
关键词
Natural language processing; Text-to-SQL; Semantic parsing; SQL; Graph neural network;
D O I
10.1109/IJCNN54540.2023.10191914
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text-to-SQL aims to parse natural language problems into SQL queries, which can provide a simple interface to access large databases enabling SQL novices a quicker entry into databases. As the Text-to-SQL field is intensively studied, more and more models use GNNs to encode heterogeneous graph information in this task, and how to better obtain path information between nodes in database schema heterogeneous graphs and question-database schema heterogeneous graphs will greatly affect the effectiveness of the model parsing. Our work intends to explore the problem of solving the encoding of heterogeneous graph meta-paths in the Text-to-SQL task. Previous approaches usually use multi-layer GNNs to aggregate topological structure information between nodes. However, they ignored the structural information embedded at the edges and also failed to obtain nodes that are not directly connected but can provide contextual information through meta-paths. To solve the above problem, we propose Multi-Hop Relational Graph Attention Network based Text-to-SQL Parsing Model (MHRGATSQL) for learning topological information between nodes while obtaining semantic information embedded in the edge topology. We use multi-hop attention to modify the relational graph attention network to diffuse the attention scores throughout the network, thus increasing the "receptive field" of each layer of RGAT. Experimental results on the large-scale cross-domain Text-to-SQL dataset Spider show that our model obtains an absolute improvement of 1.7% compared to the baseline and alleviates the over-smoothing problem in the deep network model.
引用
收藏
页数:8
相关论文
共 45 条
  • [11] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [12] LSTM: A Search Space Odyssey
    Greff, Klaus
    Srivastava, Rupesh K.
    Koutnik, Jan
    Steunebrink, Bas R.
    Schmidhuber, Juergen
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (10) : 2222 - 2232
  • [13] Guo JQ, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P4524
  • [14] Hwang W., 2019, A comprehensive exploration on wikisql with table-aware word contextualization
  • [15] Learning a Neural Semantic Parser from User Feedback
    Iyer, Srinivasan
    Konstas, Ioannis
    Cheung, Alvin
    Krishnamurthy, Jayant
    Zettlemoyer, Luke
    [J]. PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 963 - 973
  • [16] Constructing an Interactive Natural Language Interface for Relational Databases
    Li, Fei
    Jagadish, H. V.
    [J]. PROCEEDINGS OF THE VLDB ENDOWMENT, 2014, 8 (01): : 73 - 84
  • [17] Lin XV, 2020, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, P4870
  • [18] Loshchilov I., 2018, INT C LEARNING REPRE
  • [19] Oono K., 2019, INT C LEARNING REPRE
  • [20] Paszke A, 2019, ADV NEUR IN, V32