Graph Neural Networks for Natural Language Processing: A Survey

被引:121
作者
Wu, Lingfei [1 ]
Chen, Yu [2 ]
Shen, Kai [3 ,7 ]
Guo, Xiaojie [1 ]
Gao, Hanning [4 ]
Li, Shucheng [5 ]
Pei, Jian [6 ]
Long, Bo [7 ]
机构
[1] JD COM Silicon Valley Res Ctr, Mountain View, CA 94043 USA
[2] Rensselaer Polytech Inst, Troy, NY 12181 USA
[3] Zhejiang Univ, Hangzhou, Zhejiang, Peoples R China
[4] Cent China Normal Univ, Wuhan, Peoples R China
[5] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing, Peoples R China
[6] Simon Fraser Univ, Burnaby, BC, Canada
[7] JD COM, Beijing, Peoples R China
来源
FOUNDATIONS AND TRENDS IN MACHINE LEARNING | 2023年 / 16卷 / 02期
关键词
CONVOLUTIONAL NETWORKS; TRANSLATION; SEQUENCE; MEMORY;
D O I
10.1561/2200000096
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning has become the dominant approach in addressing various tasks in Natural Language Processing (NLP). Although text inputs are typically represented as a sequence of tokens, there is a rich variety of NLP problems that can be best expressed with a graph structure. As a result, there is a surge of interest in developing new deep learning techniques on graphs for a large number of NLP tasks. In this survey, we present a comprehensive overview on Graph Neural Networks (GNNs) for Natural Language Processing. We propose a new taxonomy of GNNs for NLP, which systematically organizes existing research of GNNs for NLP along three axes: graph construction, graph representation learning, and graph based encoder-decoder models. We further introduce a large number of NLP applications that exploits the power of GNNs and summarize the corresponding benchmark datasets, evaluation metrics, and open-source codes. Finally, we discuss various outstanding challenges for making the full use of GNNs for NLP as well as future research directions. To the best of our knowledge, this is the first comprehensive overview of Graph Neural Networks for Natural Language Processing.
引用
收藏
页码:119 / 329
页数:211
相关论文
共 442 条
[21]  
Bao JW, 2018, AAAI CONF ARTIF INTE, P5020
[22]  
Bastings J., 2017, P 2017 C EMP METH NA, P1957, DOI 10.18653/v1/d17- 1209
[23]  
Beck D, 2018, PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, P273
[24]  
Ben Veyseh AP, 2020, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, P4543
[25]  
Bengio S, 2015, ADV NEUR IN, V28
[26]   The Nested Chinese Restaurant Process and Bayesian Nonparametric Inference of Topic Hierarchies [J].
Blei, David M. ;
Griffiths, Thomas L. ;
Jordan, Michael I. .
JOURNAL OF THE ACM, 2010, 57 (02)
[27]   Latent Dirichlet allocation [J].
Blei, DM ;
Ng, AY ;
Jordan, MI .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (4-5) :993-1022
[28]  
Blitzer John, 2007, ACL
[29]  
Bogin B, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P3659
[30]  
Bogin B, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P4560