Understanding the black-box: towards interpretable and reliable deep learning models

被引:18
作者
Qamar, Tehreem [1 ]
Bawany, Narmeen Zakaria [1 ]
机构
[1] Jinnah Univ Women, Ctr Comp Res, Dept Comp Sci & Software Engn, Karachi, Pakistan
关键词
Deep learning; Explainable AI; Transfer learning; Pre-trained models; IMAGE CLASSIFICATION;
D O I
10.7717/peerj-cs.1629
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning (DL) has revolutionized the field of artificial intelligence by providing sophisticated models across a diverse range of applications, from image and speech recognition to natural language processing and autonomous driving. However, deep learning models are typically black-box models where the reason for predictions is unknown. Consequently, the reliability of the model becomes questionable in many circumstances. Explainable AI (XAI) plays an important role in improving the transparency and interpretability of the model thereby making it more reliable for real-time deployment. To investigate the reliability and truthfulness of DL models, this research develops image classification models using transfer learning mechanism and validates the results using XAI technique. Thus, the contribution of this research is twofold, we employ three pre-trained models VGG16, MobileNetV2 and ResNet50 using multiple transfer learning techniques for a fruit classification task consisting of 131 classes. Next, we inspect the reliability of models, based on these pre-trained networks, by utilizing Local Interpretable Model-Agnostic Explanations, the LIME, a popular XAI technique that generates explanations for the predictions. Experimental results reveal that transfer learning provides optimized results of around 98% accuracy. The classification of the models is validated on different instances using LIME and it was observed that each model predictions are interpretable and understandable as they are based on pertinent image features that are relevant to particular classes. We believe that this research gives an insight for determining how an interpretation can be drawn from a complex AI model such that its accountability and trustworthiness can be increased.
引用
收藏
页数:21
相关论文
共 53 条
[41]   "Why Should I Trust You?" Explaining the Predictions of Any Classifier [J].
Ribeiro, Marco Tulio ;
Singh, Sameer ;
Guestrin, Carlos .
KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, :1135-1144
[42]  
Sakib S, 2020, Arxiv, DOI arXiv:1904.00783
[43]  
Samek W, 2017, Arxiv, DOI arXiv:1708.08296
[44]   MobileNetV2: Inverted Residuals and Linear Bottlenecks [J].
Sandler, Mark ;
Howard, Andrew ;
Zhu, Menglong ;
Zhmoginov, Andrey ;
Chen, Liang-Chieh .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4510-4520
[45]   Deep Learning in Image Classification using Residual Network (ResNet) Variants for Detection of Colorectal Cancer [J].
Sarwinda, Devvi ;
Paradisa, Radifa Hilya ;
Bustamam, Alhadi ;
Anggia, Pinkie .
5TH INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND COMPUTATIONAL INTELLIGENCE 2020, 2021, 179 :423-431
[46]   Effectiveness of Transfer Learning and Fine Tuning in Automated Fruit Image Classification [J].
Siddiqi, Raheel .
ICDLT 2019: 2019 3RD INTERNATIONAL CONFERENCE ON DEEP LEARNING TECHNOLOGIES, 2019, :91-100
[47]  
Simonyan K, 2015, Arxiv, DOI arXiv:1409.1556
[48]  
Torrey L., 2010, Handbook of research on machine learning applications and trends: algorithms, methods, and techniques, P242, DOI DOI 10.4018/978-1-60566-766-9.CH011
[49]   Explainable image classification with evidence counterfactual [J].
Vermeire, Tom ;
Brughmans, Dieter ;
Goethals, Sofie ;
de Oliveira, Raphael Mazzine Barbossa ;
Martens, David .
PATTERN ANALYSIS AND APPLICATIONS, 2022, 25 (02) :315-335
[50]  
Wadsworth C, 2018, Arxiv, DOI arXiv:1807.00199