Comparison of an Ensemble of Machine Learning Models and the BERT Language Model for Analysis of Text Descriptions of Brain CT Reports to Determine the Presence of Intracranial

被引:0
作者
Khoruzhaya, A. N. [1 ]
Kozlov, D. V. [2 ]
Arzamasov, K. M. [3 ]
Kremneva, E. I. [4 ]
机构
[1] Moscow Dept Hlth, Sci & Pract Clin Ctr Diagnost & Telemed Technol, Dept Innovat Technol, Bldg 1,24 Petrovka St, Moscow 127051, Russia
[2] Moscow Dept Hlth, Sci & Pract Clin Ctr Diagnost & Telemed Technol, Dept Med Informat Radi & Radiogenom, Bldg 1,24 Petrovka St, Moscow 127051, Russia
[3] Moscow Dept Hlth, Sci & Pract Clin Ctr Diagnost & Telemed Technol, Dept Med Informat Radi & Radiogenom, Bldg 1,24 Petrovka St, Moscow 127051, Russia
[4] Res Ctr Neurol, Dept Innovat Technol, 80 Volokolamskoye Shosse, Moscow 125367, Russia
基金
俄罗斯科学基金会;
关键词
computed tomography; diagnostic reports; intracranial hemorrhage; natural language processing; machine learning; BERT;
D O I
10.17691/stm2024.16.1.03
中图分类号
R-3 [医学研究方法]; R3 [基础医学];
学科分类号
1001 ;
摘要
The aim of this study is to train and test an ensemble of machine learning models, as well as to compare its performance with the BERT language model pre -trained on medical data to perform simple binary classification, i.e., determine the presence/absence of the signs of intracranial hemorrhage (ICH) in brain CT reports. Materials and Methods. Seven machine learning algorithms and three text vectorization techniques were selected as models to solve the binary classification problem. These models were trained on textual data represented by 3980 brain CT reports from 56 inpatient medical facilities in Moscow. The study utilized three text vectorization techniques: bag of words, TF-IDF, and Word2Vec. The resulting data were then processed by the following machine learning algorithms: decision tree, random forest, logistic regression, nearest neighbors, support vector machines, Catboost, and XGboost. Data analysis and pre-processing were performed using NLTK (Natural Language Toolkit, version 3.6.5), libraries for character -based and statistical processing of natural language, and Scikit-learn (version 0.24.2), a library for machine learning containing tools to tackle classification challenges. MedRuBertTiny2 was taken as a BERT transformer model pre -trained on medical data. Results. Based on the training and testing outcomes from seven machine learning algorithms, the authors selected three algorithms that yielded the highest metrics (i.e. sensitivity and specificity): CatBoost, logistic regression, and nearest neighbors. The highest metrics were achieved by the bag of words technique. These algorithms were assembled into an ensemble using the stacking technique. The sensitivity and specificity for the validation dataset separated from the original sample were 0.93 and 0.90, respectively. Next, the ensemble and the BERT model were trained on an independent dataset containing 9393 textual radiology reports also divided into training and test sets. Once the ensemble was tested on this dataset, the resulting sensitivity and specificity were 0.92 and 0.90, respectively. The BERT model tested on these data demonstrated a sensitivity of 0.97 and a specificity of 0.90. Conclusion. When analyzing textual reports of brain CT scans with signs of intracranial hemorrhage, the trained ensemble demonstrated high accuracy metrics. Still, manual quality control of the results is required during its application. The pre -trained BERT transformer model, additionally trained on diagnostic textual reports, demonstrated higher accuracy metrics (p<0.05). The results show promise in terms of finding specific values for both binary classification task and in-depth analysis of unstructured medical information.
引用
收藏
页码:27 / 34
页数:8
相关论文
共 27 条
[1]  
[Anonymous], Model DmitryPogrebnoy/MedRuBertTiny2
[2]  
Bressem KK, 2023, Arxiv, DOI [arXiv:2303.08179, DOI 10.48550/ARXIV.2303.08179, 10.48550/arXiv.2303.08179]
[3]   Turkish Medical Text Classification Using BERT [J].
Celikten, Azer ;
Bulut, Hasan .
29TH IEEE CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS (SIU 2021), 2021,
[4]   The reporting quality of natural language processing studies: systematic review of studies of radiology reports [J].
Davidson, Emma M. ;
Poon, Michael T. C. ;
Casey, Arlene ;
Grivas, Andreas ;
Duma, Daniel ;
Dong, Hang ;
Suarez-Paniagua, Victor ;
Grover, Claire ;
Tobin, Richard ;
Whalley, Heather ;
Wu, Honghan ;
Alex, Beatrice ;
Whiteley, William .
BMC MEDICAL IMAGING, 2021, 21 (01)
[5]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[6]   Improving Bert-Based Model for Medical Text Classification with an Optimization Algorithm [J].
Gasmi, Karim .
ADVANCES IN COMPUTATIONAL COLLECTIVE INTELLIGENCE, ICCCI 2022, 2022, 1653 :101-111
[7]   Natural language processing of head CT reports to identify intracranial mass effect: CTIME algorithm [J].
Gordon, Alexandra June ;
Banerjee, Imon ;
Block, Jason ;
Winstead-Derlega, Christopher ;
Wilson, Jennifer G. ;
Mitarai, Tsuyoshi ;
Jarrett, Michael ;
Sanyal, Josh ;
Rubin, Daniel L. ;
Wintermark, Max ;
Kohn, Michael A. .
AMERICAN JOURNAL OF EMERGENCY MEDICINE, 2022, 51 :388-392
[8]   Machine learning in medicine: a practical introduction to natural language processing [J].
Harrison, Conrad J. ;
Sidey-Gibbons, Chris J. .
BMC MEDICAL RESEARCH METHODOLOGY, 2021, 21 (01)
[9]   Ensemble Approaches to Recognize Protected Health Information in Radiology Reports [J].
Horng, Hannah ;
Steinkamp, Jackson ;
Kahn, Charles E., Jr. ;
Cook, Tessa S. .
JOURNAL OF DIGITAL IMAGING, 2022, 35 (06) :1694-1698
[10]   Decision tree analysis in subarachnoid hemorrhage: prediction of outcome parameters during the course of aneurysmal subarachnoid hemorrhage using decision tree analysis [J].
Hostettler, Isabel Charlotte ;
Muroi, Carl ;
Richter, Johannes Konstantin ;
Schmid, Josef ;
Neidert, Marian Christoph ;
Seule, Martin ;
Boss, Oliver ;
Pangalu, Athina ;
Germans, Menno Robbert ;
Keller, Emanuela .
JOURNAL OF NEUROSURGERY, 2018, 129 (06) :1499-1510