Retrieval augmentation of large language models for lay language generation

被引:21
作者
Guo, Yue [1 ]
Qiu, Wei [2 ]
Leroy, Gondy [3 ]
Wang, Sheng [2 ]
Cohen, Trevor [1 ]
机构
[1] Univ Washington, Biomed & Hlth Informat, Seattle, WA 98195 USA
[2] Univ Washington, Paul G Allen Sch Comp Sci, Seattle, WA USA
[3] Univ Arizona, Management Informat Syst, Tucson, AZ USA
关键词
Large language models; Retrieval-augmented model; Lay language summary; Background explanation; Text generation; TEXT; KNOWLEDGE; SYSTEM;
D O I
10.1016/j.jbi.2023.104580
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The complex linguistic structures and specialized terminology of expert-authored content limit the accessibility of biomedical literature to the general public. Automated methods have the potential to render this literature more interpretable to readers with different educational backgrounds. Prior work has framed such lay language generation as a summarization or simplification task. However, adapting biomedical text for the lay public includes the additional and distinct task of background explanation: adding external content in the form of definitions, motivation, or examples to enhance comprehensibility. This task is especially challenging because the source document may not include the required background knowledge. Furthermore, background explanation capabilities have yet to be formally evaluated, and little is known about how best to enhance them. To address this problem, we introduce Retrieval-Augmented Lay Language (RALL) generation, which intuitively fits the need for external knowledge beyond that in expert-authored source documents. In addition, we introduce CELLS, the largest (63k pairs) and broadest-ranging (12 journals) parallel corpus for lay language generation. To evaluate RALL, we augmented state-of-the-art text generation models with information retrieval of either term definitions from the UMLS and Wikipedia, or embeddings of explanations from Wikipedia documents. Of these, embedding-based RALL models improved summary quality and simplicity while maintaining factual correctness, suggesting that Wikipedia is a helpful source for background explanation in this context. We also evaluated the ability of both an open-source Large Language Model (Llama 2) and a closed-source Large Language Model (GPT-4) in background explanation, with and without retrieval augmentation. Results indicate that these LLMs can generate simplified content, but that the summary quality is not ideal. Taken together, this work presents the first comprehensive study of background explanation for lay language generation, paving the path for disseminating scientific knowledge to a broader audience. Our code and data are publicly available at: https://github.com/LinguisticAnomalies/pls_retrieval.
引用
收藏
页数:17
相关论文
共 82 条
[1]  
Alambo A., 2022, Entity-driven fact-aware abstractive summarization of biomedical literature
[2]   Inter-Coder Agreement for Computational Linguistics [J].
Artstein, Ron ;
Poesio, Massimo .
COMPUTATIONAL LINGUISTICS, 2008, 34 (04) :555-596
[3]   A dataset for plain language adaptation of biomedical abstracts [J].
Attal, Kush ;
Ondov, Brian ;
Demner-Fushman, Dina .
SCIENTIFIC DATA, 2023, 10 (01) :8
[4]  
Banerjee S, 2005, P ACL WORKSH INTR EX
[5]   COVID-19 Misinformation Online and Health Literacy: A Brief Overview [J].
Bin Naeem, Salman ;
Boulos, Maged N. Kamel .
INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH, 2021, 18 (15)
[6]   The Unified Medical Language System (UMLS): integrating biomedical terminology [J].
Bodenreider, O .
NUCLEIC ACIDS RESEARCH, 2004, 32 :D267-D270
[7]   USING KINTSCH COMPUTATIONAL MODEL TO IMPROVE INSTRUCTIONAL TEXT - EFFECTS OF REPAIRING INFERENCE CALLS ON RECALL AND COGNITIVE STRUCTURES [J].
BRITTON, BK ;
GULGOZ, S .
JOURNAL OF EDUCATIONAL PSYCHOLOGY, 1991, 83 (03) :329-345
[8]  
Brown T.B., 2020, Advances in neural information processing systems
[9]   Extractive text summarization system to aid data extraction from full text in systematic review development [J].
Bui, Duy Duc An ;
Del Fiol, Guilherme ;
Hurdle, John F. ;
Jonnalagadda, Siddhartha .
JOURNAL OF BIOMEDICAL INFORMATICS, 2016, 64 :265-272
[10]  
Cachola I., 2020, Findings of EMNLP