A text-mining tool generated title-abstract screening workload savings: performance evaluation versus single-human screening

被引:12
作者
Carey, Niamh [1 ]
Harte, Marie
Mc Cullagh, Laura
机构
[1] St James Hosp, Natl Ctr Pharmacoecon, Trinity Ctr Hlth Sci, Old Stone Bldg, Dublin, Ireland
关键词
Machine learning; Text; -mining; Screening; Systematic review; Methodology; Lymphoma;
D O I
10.1016/j.jclinepi.2022.05.017
中图分类号
R19 [保健组织与事业(卫生事业管理)];
学科分类号
摘要
Background and Objectives: Text-mining tool, Abstrackr, may potentially reduce the workload burden of title and abstract screening (Stage 1), using screening prioritization and truncation. This study aimed to evaluate the performance of Abstrackr's text-mining functions ('Abstrackr-assisted screening'; screening undertaken by a single-human screener and Abstrackr) vs. Single-human screening. Methods: A systematic review of treatments for relapsed/refractory diffuse large B cell lymphoma (n = 7,723) was used. Citations, uploaded to Abstrackr, were screened by a human screener until a pre-specified maximum prediction score of 0.39540 was reached. Ab-strackr's predictions were compared with the judgments of a second, human screener (who screened all citations in Covidence). The per-formance metrics were sensitivity, specificity, precision, false negative rate, proportion of relevant citations missed, workload savings, and time savings. Results: Abstrackr reduced Stage 1 workload by 67% (5.4 days), when compared with Single-human screening. Sensitivity was high (91%). The false negative rate at Stage 1 was 9%; however, none of those citations were included following full-text screening. The high proportion of false positives (n = 2,001) resulted in low specificity (72%) and precision (15.5%). Conclusion: Abstrackr-assisted screening provided Stage 1 workload savings that did not come at the expense of omitting relevant citations. However, Abstrackr overestimated citation relevance, which may have negative workload implications at full-text screening. (C) 2022 The Author(s). Published by Elsevier Inc.
引用
收藏
页码:53 / 59
页数:7
相关论文
共 22 条
[1]   STATISTICS NOTES - DIAGNOSTIC-TESTS-1 - SENSITIVITY AND SPECIFICITY .3. [J].
ALTMAN, DG ;
BLAND, JM .
BRITISH MEDICAL JOURNAL, 1994, 308 (6943) :1552-1552
[2]  
Ananiadou S., 2006, Text Mining for Biology and Biomedicine
[3]  
[Anonymous], 2021, Cochrane Handbook for Systematic Reviews of Interventions Version 6.2
[4]  
Balke E., 2019, KTDRR CAMPBELL COLLA
[5]   Analysis of the time and workers needed to conduct systematic reviews of medical interventions using data from the PROSPERO registry [J].
Borah, Rohit ;
Brown, Andrew W. ;
Capers, Patrice L. ;
Kaiser, Kathryn A. .
BMJ OPEN, 2017, 7 (02)
[6]  
EPPI-Reviewer, MACH LEAN FUNCT EPPI
[7]   Single-reviewer abstract screening missed 13 percent of relevant studies: a crowd -based, randomized controlled trial [J].
Gartlehner, Gerald ;
Affengruber, Lisa ;
Titscher, Viktoria ;
Noel-Storr, Anna ;
Dooley, Gordon ;
Ballarini, Nicolas ;
Koenig, Franz .
JOURNAL OF CLINICAL EPIDEMIOLOGY, 2020, 121 :20-28
[8]   Decoding semi-automated title-abstract screening: findings from a convenience sample of reviews [J].
Gates, Allison ;
Gates, Michelle ;
DaRosa, Daniel ;
Elliott, Sarah A. ;
Pillay, Jennifer ;
Rahman, Sholeh ;
Ben Vandermeer ;
Hartling, Lisa .
SYSTEMATIC REVIEWS, 2020, 9 (01)
[9]   The semi-automation of title and abstract screening: a retrospective exploration of ways to leverage Abstrackr's relevance predictions in systematic and rapid reviews [J].
Gates, Allison ;
Gates, Michelle ;
Sebastianski, Meghan ;
Guitard, Samantha ;
Elliott, Sarah A. ;
Hartling, Lisa .
BMC MEDICAL RESEARCH METHODOLOGY, 2020, 20 (01)
[10]   Performance and usability of machine learning for screening in systematic reviews: a comparative evaluation of three tools [J].
Gates, Allison ;
Guitard, Samantha ;
Pillay, Jennifer ;
Elliott, Sarah A. ;
Dyson, Michele P. ;
Newton, Amanda S. ;
Hartling, Lisa .
SYSTEMATIC REVIEWS, 2019, 8 (01)