On the Intractability to Synthesize Factual Inconsistencies in Summarization

被引:0
作者
Luc, Ge [1 ]
Fan, Weisi [1 ]
Li, Miaoran [1 ]
He, Youbiao [1 ]
Yang, Yinfei [2 ]
Bao, Forrest Sheng [1 ]
机构
[1] Iowa State Univ, Ames, IA 50011 USA
[2] Apple Inc, Cupertino, CA USA
来源
FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: EACL 2024 | 2024年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Factual consistency detection has gotten significant attention for the task of abstractive summarization. Many existing works rely on synthetic training data, which may not accurately reflect or match the inconsistencies produced by summarization models. In this paper, we first systematically analyze the shortcomings of the current methods in synthesizing inconsistent summaries. Current synthesis methods may fail to produce inconsistencies of coreference errors and discourse errors, per our quantitative and qualitative study. Then, employing the parameter-efficient finetuning (PEFT) technique, we discover that a competitive factual consistency detector can be achieved using thousands of real model-generated summaries with human annotations. Our study demonstrates the importance of real machine-generated texts with human annotation in Natural Language Generatioon (NLG) evaluation as our model outperforms the SOTA on the CoGenSumm, FactCC, Frank, and SummEval datasets.
引用
收藏
页码:1026 / 1037
页数:12
相关论文
共 33 条
[1]  
Balachandran Vidhisha, 2022, P 2022 C EMP METH NA, P9818
[2]  
Bowman Samuel R., 2015, Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, P632, DOI DOI 10.18653/V1/D15-1075
[3]  
Chan Hou Pong, 2023, FINDINGS ASS COMPUTA, P6433
[4]  
Chin-Yew Lin, 2004, Text summarization branches out, P74
[5]  
Durmus Esin, 2020, P 58 ANN M ASS COMPU, P5055
[6]   SummEval: Re-evaluating Summarization Evaluation [J].
Fabbri, Alexander R. ;
Kryscinski, Wojciech ;
McCann, Bryan ;
Xiong, Caiming ;
Socher, Richard ;
Radev, Dragomir .
TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2021, 9 :391-409
[7]  
Falke T, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P2214
[8]  
Goyal T, 2020, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020
[9]  
He P., 2020, ARXIV
[10]  
Hu Edward, 2021, LoRA: Low-Rank Adaptation of Large Language Models