Development and Validation of a Literature Screening Tool: Few-Shot Learning Approach in Systematic Reviews

被引:0
作者
Wiwatthanasetthakarn, Phongphat [1 ]
Ponthongmak, Wanchana [1 ]
Looareesuwan, Panu [1 ]
Tansawet, Amarit [2 ]
Numthavaj, Pawin [1 ]
Mckay, Gareth J. [3 ]
Attia, John [4 ]
Thakkinstian, Ammarin [1 ]
机构
[1] Mahidol Univ, Ramathibodi Hosp, Fac Med, Dept Clin Epidemiol & Biostat, 4th Floor,Sukho Pl Bldg Sukhothai Rd, Bangkok 10300, Thailand
[2] Navamindradhiraj Univ, Fac Med, Vajira Hosp, Dept Res & Med Innovat, Bangkok, Thailand
[3] Queens Univ Belfast, Ctr Publ Hlth, Belfast, North Ireland
[4] Univ Newcastle, Ctr Clin Epidemiol & Biostat, Sch Med & Publ Hlth, Callaghan, NSW, Australia
关键词
few shots learning; deep learning; natural language processing; S-BERT; systematic review; study selection; sentence-bidirectional encoder representations from transformers; CONFIDENCE-INTERVALS; METAANALYSIS;
D O I
10.2196/56863
中图分类号
R19 [保健组织与事业(卫生事业管理)];
学科分类号
摘要
Background: Systematic reviews (SRs) are considered the highest level of evidence, but their rigorous literature screening process can be time-consuming and resource-intensive. This is particularly challenging given the rapid pace of medical advancements, which can quickly make SRs outdated. Few-shot learning (FSL), a machine learning approach that learns effectively from limited data, offers a potential solution to streamline this process. Sentence-bidirectional encoder representations from transformers (S-BERT) are particularly promising for identifying relevant studies with fewer examples. Objective: This study aimed to develop a model framework using FSL to efficiently screen and select relevant studies for inclusion in SRs, aiming to reduce workload while maintaining high recall rates. Methods: We developed and validated the FSL model framework using 9 previously published SR projects (2016-2018). The framework used S-BERT with titles and abstracts as input data. Key evaluation metrics, including workload reduction, cosine similarity score, and the number needed to screen at 100% recall, were estimated to determine the optimal number of eligible studies for model training. A prospective evaluation phase involving 4 ongoing SRs was then conducted. Study selection by FSL and a secondary reviewer were compared with the principal reviewer (considered the gold standard) to estimate the false negative Results: Model development suggested an optimal range of 4-12 eligible studies for FSL training. Using 4-6 eligible studies during model development resulted in similarity thresholds for 100% recall, ranging from 0.432 to 0.636, corresponding to a workload reduction of 51.11% (95% CI 46.36-55.86) to 97.67% (95% CI 96.76-98.58). The prospective evaluation of 4 SRs aimed for a 50% workload reduction, yielding numbers needed to screen 497 to 1035 out of 995 to 2070 studies. The false negative rate ranged from 1.87% to 12.20% for the FSL model and from 5% to 56.48% for the second reviewer compared with the principal reviewer. Conclusions: Our FSL framework demonstrates the potential for reducing workload in SR screening by over 50%. However, the model did not achieve 100% recall at this threshold, highlighting the potential for omitting eligible studies. Future work should focus on developing a web application to implement the FSL framework, making it accessible to researchers.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] A systematic review of few-shot learning in medical imaging
    Pachetti, Eva
    Colantonio, Sara
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2024, 156
  • [2] Fractal Few-Shot Learning
    Zhou, Fobao
    Huang, Wenkai
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (11) : 16353 - 16367
  • [3] Explore pretraining for few-shot learning
    Yan Li
    Jinjie Huang
    Multimedia Tools and Applications, 2024, 83 : 4691 - 4702
  • [4] Few-shot Learning with Prompting Methods
    2023 6TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION AND IMAGE ANALYSIS, IPRIA, 2023,
  • [5] Few-shot learning for ear recognition
    Zhang, Jie
    Yu, Wen
    Yang, Xudong
    Deng, Fang
    PROCEEDINGS OF 2019 INTERNATIONAL CONFERENCE ON IMAGE, VIDEO AND SIGNAL PROCESSING (IVSP 2019), 2019, : 50 - 54
  • [6] Explore pretraining for few-shot learning
    Li, Yan
    Huang, Jinjie
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (2) : 4691 - 4702
  • [7] Few-Shot Learning With Geometric Constraints
    Jung, Hong-Gyu
    Lee, Seong-Whan
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (11) : 4660 - 4672
  • [8] Few-Shot Learning Meets Transformer: Unified Query-Support Transformers for Few-Shot Classification
    Wang, Xixi
    Wang, Xiao
    Jiang, Bo
    Luo, Bin
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (12) : 7789 - 7802
  • [9] Adaptive Learning Knowledge Networks for Few-Shot Learning
    Yan, Minghao
    IEEE ACCESS, 2019, 7 : 119041 - 119051
  • [10] Learning to Capture the Query Distribution for Few-Shot Learning
    Chi, Ziqiu
    Wang, Zhe
    Yang, Mengping
    Li, Dongdong
    Du, Wenli
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (07) : 4163 - 4173