Graph Neural Networks for Natural Language Processing: A Survey

被引:121
作者
Wu, Lingfei [1 ]
Chen, Yu [2 ]
Shen, Kai [3 ,7 ]
Guo, Xiaojie [1 ]
Gao, Hanning [4 ]
Li, Shucheng [5 ]
Pei, Jian [6 ]
Long, Bo [7 ]
机构
[1] JD COM Silicon Valley Res Ctr, Mountain View, CA 94043 USA
[2] Rensselaer Polytech Inst, Troy, NY 12181 USA
[3] Zhejiang Univ, Hangzhou, Zhejiang, Peoples R China
[4] Cent China Normal Univ, Wuhan, Peoples R China
[5] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing, Peoples R China
[6] Simon Fraser Univ, Burnaby, BC, Canada
[7] JD COM, Beijing, Peoples R China
来源
FOUNDATIONS AND TRENDS IN MACHINE LEARNING | 2023年 / 16卷 / 02期
关键词
CONVOLUTIONAL NETWORKS; TRANSLATION; SEQUENCE; MEMORY;
D O I
10.1561/2200000096
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning has become the dominant approach in addressing various tasks in Natural Language Processing (NLP). Although text inputs are typically represented as a sequence of tokens, there is a rich variety of NLP problems that can be best expressed with a graph structure. As a result, there is a surge of interest in developing new deep learning techniques on graphs for a large number of NLP tasks. In this survey, we present a comprehensive overview on Graph Neural Networks (GNNs) for Natural Language Processing. We propose a new taxonomy of GNNs for NLP, which systematically organizes existing research of GNNs for NLP along three axes: graph construction, graph representation learning, and graph based encoder-decoder models. We further introduce a large number of NLP applications that exploits the power of GNNs and summarize the corresponding benchmark datasets, evaluation metrics, and open-source codes. Finally, we discuss various outstanding challenges for making the full use of GNNs for NLP as well as future research directions. To the best of our knowledge, this is the first comprehensive overview of Graph Neural Networks for Natural Language Processing.
引用
收藏
页码:119 / 329
页数:211
相关论文
共 442 条
[1]   Community detection and stochastic block models: Recent developments [J].
Abbe, Emmanuel .
Journal of Machine Learning Research, 2018, 18 :1-86
[2]  
Allahyari M, 2017, Arxiv, DOI arXiv:1707.02268
[3]  
Allamanis Miltiadis, 2018, Learning to represent programs with graphs
[4]  
Alon U., 2018, code2seq: Generating sequences from structured representations of code
[5]  
Alvarez-Melis D., 2016, Tree-structured decoding with doubly-recurrent neural networks
[6]  
Amini A, 2019, Arxiv, DOI arXiv:1905.13319
[7]  
Angeli G, 2015, PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1, P344
[8]  
[Anonymous], 2016, Trans. Assoc. Comput. Linguist, DOI [DOI 10.1162/TACL_A_00101, DOI 10.1162/TACLA00101]
[9]  
[Anonymous], 2013, Advances in neural information processing systems
[10]  
[Anonymous], 2017, Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing