JointGT: Graph-Text Joint Representation Learning for Text Generation from Knowledge Graphs

被引:0
作者
Ke, Pei [1 ]
Ji, Haozhe [1 ]
Ran, Yu [2 ]
Cui, Xin [2 ]
Wang, Liwei [3 ]
Song, Linfeng [4 ]
Zhu, Xiaoyan [1 ]
Huang, Minlie [1 ]
机构
[1] Tsinghua Univ, Beijing Natl Res Ctr Informat Sci & Technol, Inst Artificial Intelligence, CoAI Grp,Dept Comp Sci & Technol, Beijing 100084, Peoples R China
[2] Sogou Inc, Beijing, Peoples R China
[3] Chinese Univ Hong Kong, Hong Kong, Peoples R China
[4] Tencent AI Lab, Shenzhen, Peoples R China
来源
FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021 | 2021年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing pre-trained models for knowledge-graph-to-text (KG-to-text) generation simply fine-tune text-to-text pre-trained models such as BART or T5 on KG-to-text datasets, which largely ignore the graph structure during encoding and lack elaborate pre-training tasks to explicitly model graph-text alignments. To tackle these problems, we propose a graph-text joint representation learning model called JointGT. During encoding, we devise a structure-aware semantic aggregation module which is plugged into each Transformer layer to preserve the graph structure. Furthermore, we propose three new pre-training tasks to explicitly enhance the graph-text alignment including respective text / graph reconstruction, and graph-text alignment in the embedding space via Optimal Transport. Experiments show that JointGT obtains new state-of-the-art performance on various KG-to-text datasets.
引用
收藏
页码:2526 / 2538
页数:13
相关论文
共 51 条
[1]  
Banerjee S, 2005, P ACL WORKSH INTR EX, P65
[2]  
Bordes A., 2013, ADV NEURAL INFORM PR, V26, P2787, DOI DOI 10.5555/2999792.2999923
[3]  
Chen L., 2020, ICML, P1542
[4]  
Chen WH, 2020, PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), P8635
[5]  
Chen Yu, 2020, ARXIV200406015
[6]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[7]  
Ferreira TC, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P552
[8]  
FLEISS JL, 1971, PSYCHOL BULL, V76, P378, DOI 10.1037/h0031619
[9]  
Fu ZH, 2020, PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), P9183
[10]  
Gardent Claire, 2017, P 10 INT C NATURAL L, P124, DOI [10.18653/v1/w17-3518, DOI 10.18653/V1/W17-3518]