Explainable machine learning in image classification models: An uncertainty quantification perspective

被引:39
作者
Zhang, Xiaoge [1 ]
Mahadevan, Sankaran [3 ]
Chan, Felix T. S. [2 ]
机构
[1] Hong Kong Polytech Univ, Dept Ind & Syst Engn, Hung Hom, Hong Kong, Peoples R China
[2] Macau Univ Sci & Technol, Dept Decis Sci, Ave Wai Long, Taipa, Macau, Peoples R China
[3] Vanderbilt Univ, Sch Engn, Dept Civil & Environm Engn, Nashville, TN 37235 USA
关键词
Deep learning; Model explainability; Uncertainty quantification; Bayesian neural network; Prediction difference analysis; NETWORK;
D O I
10.1016/j.knosys.2022.108418
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The poor explainability of deep learning models has hindered their adoption in safety and quality critical applications. This paper focuses on image classification models and aims to enhance the explainability of deep learning models through the development of an uncertainty quantification based framework. The proposed methodology consists of three major steps. In the first step, we adopt dropout-based Bayesian neural network to characterize the structure and parameter uncertainty inherent in deep learning models, propagate and represent such uncertainties to the model prediction as a distribution. Next, we employ entropy as a quantitative indicator to measure the uncertainty in model prediction, and develop an Empirical Cumulative Distribution Function (ECDF)-based approach to determine an appropriate threshold value for the purpose of deciding when to accept or reject the model prediction. Secondly, in the cases with high model prediction uncertainty, we combine the prediction difference analysis (PDA) approach with dropout-based Bayesian neural network to quantify the uncertainty in pixel-wise feature importance, and identify the locations in the input image that highly correlate with the model prediction uncertainty. In the third step, we develop a robustness-based design optimization formulation to enhance the relevance between input features and model prediction, and leverage a differential evolution approach to optimize the pixels in the input image with high uncertainty in feature importance. Experimental studies in MNIST and CIFAR10 image classifications are included to demonstrate the effectiveness of the proposed approach in increasing the explainability of deep learning models. (c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页数:17
相关论文
共 68 条
[1]   A review of uncertainty quantification in deep learning: Techniques, applications and challenges [J].
Abdar, Moloud ;
Pourpanah, Farhad ;
Hussain, Sadiq ;
Rezazadegan, Dana ;
Liu, Li ;
Ghavamzadeh, Mohammad ;
Fieguth, Paul ;
Cao, Xiaochun ;
Khosravi, Abbas ;
Acharya, U. Rajendra ;
Makarenkov, Vladimir ;
Nahavandi, Saeid .
INFORMATION FUSION, 2021, 76 :243-297
[2]   A survey of network anomaly detection techniques [J].
Ahmed, Mohiuddin ;
Mahmood, Abdun Naser ;
Hu, Jiankun .
JOURNAL OF NETWORK AND COMPUTER APPLICATIONS, 2016, 60 :19-31
[3]   Deep learning approaches for anomaly-based intrusion detection systems: A survey, taxonomy, and open issues [J].
Aldweesh, Arwa ;
Derhab, Abdelouahid ;
Emam, Ahmed Z. .
KNOWLEDGE-BASED SYSTEMS, 2020, 189
[4]   Image classification with deep learning in the presence of noisy labels: A survey [J].
Algan, Gorkem ;
Ulusoy, Ilkay .
KNOWLEDGE-BASED SYSTEMS, 2021, 215
[5]  
Alvarez-Melis D, 2018, ADV NEUR IN, V31
[6]  
[Anonymous], 2018, INT C MED IM DEEP LE
[7]   On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation [J].
Bach, Sebastian ;
Binder, Alexander ;
Montavon, Gregoire ;
Klauschen, Frederick ;
Mueller, Klaus-Robert ;
Samek, Wojciech .
PLOS ONE, 2015, 10 (07)
[8]  
Blundell C, 2015, PR MACH LEARN RES, V37, P1613
[9]  
Burkart N, 2021, J ARTIF INTELL RES, V70, P245
[10]  
Chai Lucy R, 2018, Uncertainty Estimation in Bayesian Neural Networks and Links to Interpretability