Entity-Driven Fact-Aware Abstractive Summarization of Biomedical Literature

被引:4
作者
Alambo, Amanuel [1 ]
Banerjee, Tanvi [1 ]
Thirunarayan, Krishnaprasad [1 ]
Raymer, Michael [1 ]
机构
[1] Wright State Univ, Dayton, OH 45435 USA
来源
2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR) | 2022年
关键词
Transformers; Named Entity Recognition; Knowledge Bases; Abstractive Summarization; ICD-11; Knowledge Retrieval; Knowledge-enhanced Natural Language Generation; SYSTEM;
D O I
10.1109/ICPR56361.2022.9956656
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As part of the large number of scientific articles being published every year, the publication rate of biomedical literature has been increasing. Consequently, there has been considerable effort to harness and summarize the massive amount of biomedical research articles. While transformer-based encoder-decoder models in a vanilla source document-to-summary setting have been extensively studied for abstractive summarization in different domains, their major limitations continue to be entity hallucination (a phenomenon where generated summaries constitute entities not related to or present in source article(s)) and factual inconsistency. This problem is exacerbated in a biomedical setting where named entities and their semantics (which can be captured through a knowledge base) constitute the essence of an article. The use of named entities and facts mined from background knowledge bases pertaining to the named entities to guide abstractive summarization has not been studied in biomedical article summarization literature. In this paper, we propose an entity-driven fact-aware framework for training end-to-end transformer-based encoder-decoder models for abstractive summarization of biomedical articles. We call the proposed approach, whose building block is a transformer-based model, EFAS, Entity-driven Fact-aware Abstractive Summarization. We conduct a set of experiments using five state-of-the-art transformer-based encoder-decoder models (two of which are specifically designed for long document summarization) and demonstrate that injecting knowledge into the training/inference phase of these models enables the models to achieve significantly better performance than the standard source document-to-summary setting in terms of entity-level factual accuracy, N-gram novelty, and semantic equivalence while performing comparably on ROUGE metrics. The proposed approach is evaluated on ICD-11-Summ-1000, a dataset we build for abstractive summarization of biomedical literature, and PubMed-50k, a segment of a large-scale benchmark dataset for abstractive summarization of biomedical literature.
引用
收藏
页码:613 / 620
页数:8
相关论文
共 72 条
[1]   Clinical Context-Aware Biomedical Text Summarization Using Deep Neural Network: Model Development and Validation [J].
Afzal, Muhammad ;
Alam, Fakhare ;
Malik, Khalid Mahmood ;
Malik, Ghaus M. .
JOURNAL OF MEDICAL INTERNET RESEARCH, 2020, 22 (10)
[2]  
An C, 2021, arXiv
[3]  
Banerjee S, 2015, PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), P1208
[4]  
Beltagy I, 2020, Arxiv, DOI [arXiv:2004.05150, 10.48550/arXiv.2004.05150]
[5]   The Unified Medical Language System (UMLS): integrating biomedical terminology [J].
Bodenreider, O .
NUCLEIC ACIDS RESEARCH, 2004, 32 :D267-D270
[6]  
Cachola I, 2020, Arxiv, DOI [arXiv:2004.15011, DOI 10.48550/ARXIV.2004.15011]
[7]   Learning a similarity metric discriminatively, with application to face verification [J].
Chopra, S ;
Hadsell, R ;
LeCun, Y .
2005 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 1, PROCEEDINGS, 2005, :539-546
[8]  
Chopra S., 2016, 15 ANN C N AM CHAPT, P93
[9]  
Cohan A, 2018, Arxiv, DOI arXiv:1804.05685
[10]  
DeYoung J., 2021, arXiv