RST Discourse Parsing as Text-to-Text Generation

被引:2
作者
Hu, Xinyu [1 ]
Wan, Xiaojun [1 ]
机构
[1] Peking Univ, MOE Key Lab Computat Linguist, Wangxuan Inst Comp Technol, Beijing 100871, Peoples R China
基金
美国国家科学基金会;
关键词
Discourse parsing; rhetorical structure theory; text generation;
D O I
10.1109/TASLP.2023.3306710
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Previous studies have made great advances in RST discourse parsing through specific neural frameworks or features, but they usually split the parsing process into two subtasks and heavily depended on gold discourse segmentation. In this article, we introduce an end-to-end method for sentence-level RST discourse parsing via transforming it into a text-to-text generation task, which can also be simply applied to document-level parsing. Our method unifies the traditional two-stage parsing and generates the parsing tree directly from the input text through our constrained decoding and postprocessing algorithms, without requiring a complicated model. Moreover, the discourse segmentation can be simultaneously generated and extracted from the parsing tree. Experimental results on the RST Discourse Treebank demonstrate that our proposed method outperforms existing methods in both the tasks of discourse parsing and segmentation. We further carry out ablation studies and more targeted comparisons with traditional patterns to analyze our method in more detail. Considering the lack of annotated data in RST parsing, we also create high-quality augmented data and implement self-training, which further improves the performance of our method.
引用
收藏
页码:3278 / 3289
页数:12
相关论文
共 57 条
  • [1] [Anonymous], 2014, P 2014 C EMP METH NA
  • [2] [Anonymous], 2010, Dialogue and Discourse, DOI DOI 10.5087/DAD.2010.003
  • [3] Bhatia P, 2015, Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, P2212
  • [4] Braud C., 2016, P INT C COMP LING, P1903
  • [5] Carlson L., 2001, PROC ANN M SPECIAL I, P1
  • [6] Carlson L., 2001, ISI Tech. Rep. ISI-TR-545
  • [7] Chen PZ, 2020, 58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), P1672
  • [8] Clark K., 2020, PROC INT C LEARN REP
  • [9] duVerle D., 2009, P ANN M ASS COMP LIN, P665
  • [10] Feng VW, 2014, PROCEEDINGS OF THE 52ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, P511