A data interpretation approach for deep learning-based prediction models

被引:3
作者
Dadsetan, Saba [1 ]
Wu, Shandong [1 ,2 ,3 ,4 ]
机构
[1] Univ Pittsburgh, Sch Comp & Informat, Intelligent Syst Program, 210 S Bouquet St, Pittsburgh, PA 15213 USA
[2] Univ Pittsburgh, Dept Radiol, 3362 Fifth Ave, Pittsburgh, PA 15213 USA
[3] Univ Pittsburgh, Dept Biomed Informat, 3362 Fifth Ave, Pittsburgh, PA 15213 USA
[4] Univ Pittsburgh, Dept Bioengn, 3362 Fifth Ave, Pittsburgh, PA 15213 USA
来源
MEDICAL IMAGING 2019: IMAGING INFORMATICS FOR HEALTHCARE, RESEARCH, AND APPLICATIONS | 2019年 / 10954卷
基金
美国国家卫生研究院;
关键词
Deep learning; model interpretation; data interpretation; breast cancer; contribution map;
D O I
10.1117/12.2513098
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Deep learning models based on Convolutional Neural Networks (CNN) are known as successful tools in many classification and segmentation studies. Although these kinds of tools can achieve impressive performance, we still lack effective means to interpret the models, features, and the associated input data on how a model can work well in a data-driven manner. In this paper, we propose a novel investigation to interpret a deep-learning-based model for breast cancer risk prediction using screening digital mammogram images. First, we build a CNN-based risk prediction model by using normal screening mammogram images. Then we developed two different/separate schemes to explore the interpretability. In Scheme 1, we apply a sliding window-based approach to modify the input images; that is, we only keep the sub-regional imaging data inside the sliding window but padding other regions with zeros, and we observe how such an effective sub-regional input may lead to changes in the model's performance. We generated heatmaps of the AUCs with regards to all sliding windows and showed that the heatmaps can help interpret a potential correlation/response between given sliding windows and the model AUC variation. In Scheme 2, we followed the saliency map-based approach to create a Contribution Map (CM), where the CM value of each pixel reflects the strength of that pixels contributes to the prediction of the output label. Then over a CM, we identify a bounding box around the most informative sub-area of a CM to interpret the corresponding sub-area in the images as the region that is most predictive of the risk. This preliminary study demonstrates a proof of concept on developing an effective means to interpret deep learning CNN models.
引用
收藏
页数:7
相关论文
共 11 条
[1]  
Aboutalib S. S., 2018, INT SOC OPTICS PHOTO, V10575
[2]   Automatic breast density classification using neural network [J].
Arefan, D. ;
Talebpour, A. ;
Ahmadinejhad, N. ;
Asl, Kamali .
JOURNAL OF INSTRUMENTATION, 2015, 10
[3]  
Arefan D., 2013, RES J APPL SCI ENG T, V5, P513
[4]  
Erhan D., 2009, U MONTREAL, V1341, P1, DOI DOI 10.2464/JILM.23.425
[5]  
He K., 2016, CVPR, DOI [10.1109/CVPR.2016.90, DOI 10.1109/CVPR.2016.90]
[6]  
Liang HL, 2017, ASIA CONTROL CONF AS, P2340, DOI 10.1109/ASCC.2017.8287540
[7]  
Luo Y, 2011, LECT NOTES COMPUT SC, V6494, P396, DOI 10.1007/978-3-642-19318-7_31
[8]  
Simonyan K., 2013, ABS13126034 CORR
[9]  
Xie DF, 2017, APPL COMPUT INTELL S, V2017, DOI 10.1155/2017/1320780
[10]   Visualizing and Understanding Convolutional Networks [J].
Zeiler, Matthew D. ;
Fergus, Rob .
COMPUTER VISION - ECCV 2014, PT I, 2014, 8689 :818-833