PROMETEO: A CNN-Based Computer-Aided Diagnosis System for WSI Prostate Cancer Detection

被引:46
作者
Duran-Lopez, Lourdes [1 ]
Dominguez-Morales, Juan P. [1 ]
Felix Conde-Martin, Antonio [2 ]
Vicente-Diaz, Saturnino [1 ]
Linares-Barranco, Alejandro [1 ]
机构
[1] Univ Seville, Robot & Technol Comp Lab, Seville 41012, Spain
[2] Virgen de Valme Hosp, Pathol Anat Unit, Seville 41014, Spain
关键词
Convolutional neural networks; computer-aided diagnosis; deep learning; medical image analysis; prostate cancer; whole-slide images; BIOPSIES; CLASSIFICATION; NORMALIZATION;
D O I
10.1109/ACCESS.2020.3008868
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Prostate cancer is currently one of the most commonly-diagnosed types of cancer among males. Although its death rate has dropped in the last decades, it is still a major concern and one of the leading causes of cancer death. Prostate biopsy is a test that confirms or excludes the presence of cancer in the tissue. Samples extracted from biopsies are processed and digitized, obtaining gigapixel-resolution images called whole-slide images, which are analyzed by pathologists. Automated intelligent systems could be useful for helping pathologists in this analysis, reducing fatigue and making the routine process faster. In this work, a novel Deep Learning based computer-aided diagnosis system is presented. This system is able to analyze whole-slide histology images that are first patch-sampled and preprocessed using different filters, including a novel patch-scoring algorithm that removes worthless areas from the tissue. Then, patches are used as input to a custom Convolutional Neural Network, which gives a report showing malignant regions on a heatmap. The impact of applying a stain-normalization process to the patches is also analyzed in order to reduce color variability between different scanners. After training the network with a 3-fold cross-validation method, 99.98% accuracy, 99.98% F1 score and 0.999 AUC are achieved on a separate test set. The computation time needed to obtain the heatmap of a whole-slide image is, on average, around 15 s. Our custom network outperforms other state-of-the-art works in terms of computational complexity for a binary classification task between normal and malignant prostate whole-slide images at patch level.
引用
收藏
页码:128613 / 128628
页数:16
相关论文
共 35 条
  • [1] [Anonymous], 2009, Proceedings of the Optical Tissue Image analysis in Microscopy, Histopathology and Endoscopy (MICCAI Workshop)
  • [2] [Anonymous], P SPIE
  • [3] [Anonymous], 2017, P SPIE MED IMAG
  • [4] Automated Gleason grading of prostate cancer tissue microarrays via deep learning
    Arvaniti, Eirini
    Fricker, Kim S.
    Moret, Michael
    Rupp, Niels
    Hermanns, Thomas
    Fankhauser, Christian
    Wey, Norbert
    Wild, Peter J.
    Ruschoff, Jan H.
    Claassen, Manfred
    [J]. SCIENTIFIC REPORTS, 2018, 8
  • [5] Prostate needle biopsies: interobserver variation and clinical consequences of histopathological re-evaluation
    Berg, Kasper Drimer
    Toft, Birgitte Gronkaer
    Roder, Martin Andreas
    Brasso, Klaus
    Vainer, Ben
    Iversen, Peter
    [J]. APMIS, 2011, 119 (4-5) : 239 - 246
  • [6] Automated deep-learning system for Gleason grading of prostate cancer using biopsies: a diagnostic study
    Bulten, Wouter
    Pinckaers, Hans
    van Boven, Hester
    Vink, Robert
    de Bel, Thomas
    van Ginneken, Bram
    van der Laak, Jeroen
    Hulsbergen-van de Kaa, Christina
    Litjens, Geert
    [J]. LANCET ONCOLOGY, 2020, 21 (02) : 233 - 241
  • [7] Clinical-grade computational pathology using weakly supervised deep learning on whole slide images
    Campanella, Gabriele
    Hanna, Matthew G.
    Geneslaw, Luke
    Miraflor, Allen
    Silva, Vitor Werneck Krauss
    Busam, Klaus J.
    Brogi, Edi
    Reuter, Victor E.
    Klimstra, David S.
    Fuchs, Thomas J.
    [J]. NATURE MEDICINE, 2019, 25 (08) : 1301 - +
  • [8] Canziani A., 2016, ARXIV
  • [9] Ciompi F, 2017, I S BIOMED IMAGING, P160, DOI 10.1109/ISBI.2017.7950492
  • [10] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848