Reproducibility of artificial intelligence models in computed tomography of the head: a quantitative analysis

被引:0
作者
Felix Gunzer
Michael Jantscher
Eva M. Hassler
Thomas Kau
Gernot Reishofer
机构
[1] Medical University Graz,Division of Neuroradiology, Vascular and Interventional Radiology, Department of Radiology
[2] Know-Center GmbH,Research Center for Data
[3] Landeskrankenhaus Villach,Driven Business Big Data Analytics
[4] Medical University Graz,Department of Radiology
[5] BioTechMed Graz,Department of Radiology
来源
Insights into Imaging | / 13卷
关键词
Artificial intelligence; Head CT; Reproducibility; Epidemiology; Machine learning;
D O I
暂无
中图分类号
学科分类号
摘要
When developing artificial intelligence (AI) software for applications in radiology, the underlying research must be transferable to other real-world problems. To verify to what degree this is true, we reviewed research on AI algorithms for computed tomography of the head. A systematic review was conducted according to the preferred reporting items for systematic reviews and meta-analyses. We identified 83 articles and analyzed them in terms of transparency of data and code, pre-processing, type of algorithm, architecture, hyperparameter, performance measure, and balancing of dataset in relation to epidemiology. We also classified all articles by their main functionality (classification, detection, segmentation, prediction, triage, image reconstruction, image registration, fusion of imaging modalities). We found that only a minority of authors provided open source code (10.15%, n 0 7), making the replication of results difficult. Convolutional neural networks were predominantly used (32.61%, n = 15), whereas hyperparameters were less frequently reported (32.61%, n = 15). Data sets were mostly from single center sources (84.05%, n = 58), increasing the susceptibility of the models to bias, which increases the error rate of the models. The prevalence of brain lesions in the training (0.49 ± 0.30) and testing (0.45 ± 0.29) datasets differed from real-world epidemiology (0.21 ± 0.28), which may overestimate performances. This review highlights the need for open source code, external validation, and consideration of disease prevalence.
引用
收藏
相关论文
共 42 条
  • [11] Wang W(2020)Extensions of the external validation for checking learned model interpretability and generalizability Patterns 290 590-undefined
  • [12] Kiik M(2019)Convolutional neural networks for radiologic images: a radiologist’s guide Radiology 31 198-undefined
  • [13] Peek N(2007)Computer-aided diagnosis in medical imaging: historical review, current status and future potential Comput Med Imaging Graph 322 843-undefined
  • [14] Mongan J(2019)Trends in use of medical imaging in US health care systems and in Ontario, Canada, 2000–2016 JAMA undefined undefined-undefined
  • [15] Moy L(undefined)undefined undefined undefined undefined-undefined
  • [16] Kahn CE(undefined)undefined undefined undefined undefined-undefined
  • [17] Plesser HE(undefined)undefined undefined undefined undefined-undefined
  • [18] Park SH(undefined)undefined undefined undefined undefined-undefined
  • [19] Han K(undefined)undefined undefined undefined undefined-undefined
  • [20] Page MJ(undefined)undefined undefined undefined undefined-undefined