Enhancing Abstractive Summarization with Extracted Knowledge Graphs and Multi-Source Transformers

被引:20
作者
Chen, Tong [1 ]
Wang, Xuewei [1 ]
Yue, Tianwei [1 ]
Bai, Xiaoyu [2 ]
Le, Cindy X. X. [3 ]
Wang, Wenping [1 ]
机构
[1] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
[2] Rice Univ, Houston, TX 77005 USA
[3] Columbia Univ, New York, NY 10027 USA
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 13期
关键词
abstractive summarization; knowledge graph; multi-source transformers; pre-trained language models;
D O I
10.3390/app13137753
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
As the popularity of large language models (LLMs) has risen over the course of the last year, led by GPT-3/4 and especially its productization as ChatGPT, we have witnessed the extensive application of LLMs to text summarization. However, LLMs do not intrinsically have the power to verify the correctness of the information they supply and generate. This research introduces a novel approach to abstractive summarization, aiming to address the limitations of LLMs in that they struggle to understand the truth. The proposed method leverages extracted knowledge graph information and structured semantics as a guide for summarization. Building upon BART, one of the state-of-the-art sequence-to-sequence pre-trained LLMs, multi-source transformer modules are developed as an encoder, which are capable of processing textual and graphical inputs. Decoding is performed based on this enriched encoding to enhance the summary quality. The Wiki-Sum dataset, derived from Wikipedia text dumps, is introduced for evaluation purposes. Comparative experiments with baseline models demonstrate the strengths of the proposed approach in generating informative and relevant summaries. We conclude by presenting our insights into utilizing LLMs with graph external information, which will become a powerful aid towards the goal of factually correct and verified LLMs.
引用
收藏
页数:14
相关论文
共 30 条
[1]  
Angeli G, 2015, PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1, P344
[2]  
[Anonymous], 2023, Gpt-4 technical report
[3]  
Beck D, 2018, PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, P273
[4]  
Bordes A., 2013, P 26 INT C NEURAL IN, P2787
[5]  
Cao ZQ, 2018, AAAI CONF ARTIF INTE, P4784
[6]  
Damonte M, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P3649
[7]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[8]  
Dhingra B, 2020, Arxiv, DOI arXiv:2002.10640
[9]  
Dongwook Lee, 2019, arXiv
[10]  
Dou ZY, 2021, 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), P4830