Improving AMR Parsing with Sequence-to-Sequence Pre-training

被引:0
作者
Xu, Dongqin [1 ]
Li, Junhui [1 ]
Zhu, Muhua [2 ]
Min Zhang [1 ]
Zhou, Guodong [1 ]
机构
[1] Soochow Univ, Sch Comp Sci & Technol, Suzhou, Peoples R China
[2] Tencent News, Beijing, Peoples R China
来源
PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP) | 2020年
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the literature, the research on abstract meaning representation (AMR) parsing is much restricted by the size of human-curated dataset which is critical to build an AMR parser with good performance. To alleviate such data size restriction, pre-trained models have been drawing more and more attention in AMR parsing. However, previous pre-trained models, like BERT, are implemented for general purpose which may not work as expected for the specific task of AMR parsing. In this paper, we focus on sequence-to-sequence (seq2seq) AMR parsing and propose a seq2seq pre-training approach to build pre-trained models in both single and joint way on three relevant tasks, i.e., machine translation, syntactic parsing, and AMR parsing itself. Moreover, we extend the vanilla fine-tuning method to a multi-task learning fine-tuning method that optimizes for the performance of AMR parsing while endeavors to preserve the response of pre-trained models. Extensive experimental results on two English benchmark datasets show that both the single and joint pre-trained models significantly improve the performance (e.g., from 71.5 to 80.2 on AMR 2.0), which reaches the state of the art. The result is very encouraging since we achieve this with seq2seq models rather than complex models. We make our code and model available at https:// github.com/xdqkid/S2S- AMR- Parser.
引用
收藏
页码:2501 / 2511
页数:11
相关论文
共 47 条
[1]  
[Anonymous], 2017, Transactions of the Association for Computational Linguistics, DOI [DOI 10.1162/TACL_A_00065, 10.1162/tacl_a_00065]
[2]  
[Anonymous], 2016, P 2016 C EMPIRICAL M
[3]  
[Anonymous], 2017, Computational Linguistics in the Netherlands Journal, V7, P93
[4]  
[Anonymous], 2019, COMPUTING RES REPOSI
[5]  
[Anonymous], 2016, P 2016 C EMP METH NA, DOI DOI 10.18653/V1/D16-1159
[6]  
Banarescu L., 2013, P 7 LINGUISTIC ANNOT
[7]  
Cai D, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P3799
[8]  
Cai D, 2020, 58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), P1290
[9]  
Cai Shu, 2013, Short Papers, P748
[10]  
Dai AM, 2015, ADV NEUR IN, V28