Improved Fine-Tuning of In-Domain Transformer Model for Inferring COVID-19 Presence in Multi-Institutional Radiology Reports

被引:12
作者
Chambon, Pierre [1 ]
Cook, Tessa S. [2 ]
Langlotz, Curtis P. [3 ]
机构
[1] Stanford Univ, Paris Saclay Univ, Ecole Cent Paris, Stanford, CA 94305 USA
[2] Univ Penn, Philadelphia, PA 19104 USA
[3] Stanford Univ, Stanford, CA 94305 USA
基金
美国国家卫生研究院;
关键词
Radiology; COVID-19; Classification; Natural language processing (NLP); Transformer; BERT;
D O I
10.1007/s10278-022-00714-8
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Building a document-level classifier for COVID-19 on radiology reports could help assist providers in their daily clinical routine, as well as create large numbers of labels for computer vision models. We have developed such a classifier by fine-tuning a BERT-like model initialized from RadBERT, its continuous pre-training on radiology reports that can be used on all radiology-related tasks. RadBERT outperforms all biomedical pre-trainings on this COVID-19 task (P<0.01) and helps our fine-tuned model achieve an 88.9 macro-averaged F1-score, when evaluated on both X-ray and CT reports. To build this model, we rely on a multi-institutional dataset re-sampled and enriched with concurrent lung diseases, helping the model to resist to distribution shifts. In addition, we explore a variety of fine-tuning and hyperparameter optimization techniques that accelerate fine-tuning convergence, stabilize performance, and improve accuracy, especially when data or computational resources are limited. Finally, we provide a set of visualization tools and explainability methods to better understand the performance of the model, and support its practical use in the clinical setting. Our approach offers a ready-to-use COVID-19 classifier and can be applied similarly to other radiology report classification tasks.
引用
收藏
页码:164 / 177
页数:14
相关论文
共 40 条
[1]  
Abacha Asma Ben, 2021, P 20 WORKSH BIOM LAN, P74
[2]  
Alsentzer Emily., 2019, Proceedings of the 2nd Clinical Natural Language Processing Workshop, P72, DOI [DOI 10.18653/V1/W19-1909, 10.18653/v1/W19-1909]
[3]  
Beltagy I, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P3615
[4]  
Bergstra J.S., 2011, ADV NEURAL INFORM PR
[5]  
Biewald L., 2020, Experiment Tracking with Weights and Biases
[6]  
CCRB, 2022, SAMPLE SIZE CALCULAT
[7]  
CDC, 2022, CDC COV DAT TRACK
[8]  
[Clark Kevin ELECTRA ELECTRA], 2020, arXiv, DOI [DOI 10.48550/ARXIV.2003.10555, DOI 10.48550/arXiv.2003.10555, 10.48550/arXiv.2003.10555]
[9]   Bringing Transparency Design into Practice [J].
Eiband, Malin ;
Schneider, Hanna ;
Bilandzic, Mark ;
Fazekas-Con, Julian ;
Haug, Mareike ;
Hussmann, Heinrich .
IUI 2018: PROCEEDINGS OF THE 23RD INTERNATIONAL CONFERENCE ON INTELLIGENT USER INTERFACES, 2018, :211-223
[10]  
Falkner S, 2018, PR MACH LEARN RES, V80