Right for the Wrong Reason: Can Interpretable ML Techniques Detect Spurious Correlations?

被引:4
作者
Sun, Susu [1 ]
Koch, Lisa M. [2 ,3 ]
Baumgartner, Christian F. [1 ]
机构
[1] Univ Tubingen, Cluster Excellence ML Sci, Tubingen, Germany
[2] Univ Tubingen, Hertie Inst AI Brain Hlth, Tubingen, Germany
[3] Univ Tubingen, Inst Ophthalm Res, Tubingen, Germany
来源
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT II | 2023年 / 14221卷
关键词
Interpretable machine learning; Confounder detection;
D O I
10.1007/978-3-031-43895-0_40
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While deep neural network models offer unmatched classification performance, they are prone to learning spurious correlations in the data. Such dependencies on confounding information can be difficult to detect using performance metrics if the test data comes from the same distribution as the training data. Interpretable ML methods such as post-hoc explanations or inherently interpretable classifiers promise to identify faulty model reasoning. However, there is mixed evidence whether many of these techniques are actually able to do so. In this paper, we propose a rigorous evaluation strategy to assess an explanation technique's ability to correctly identify spurious correlations. Using this strategy, we evaluate five post-hoc explanation techniques and one inherently interpretable method for their ability to detect three types of artificially added confounders in a chest x-ray diagnosis task. We find that the post-hoc technique SHAP, as well as the inherently interpretable Attri-Net provide the best performance and can be used to reliably identify faulty model behavior.
引用
收藏
页码:425 / 434
页数:10
相关论文
共 27 条
[1]  
Adebayo J., 2022, INT C LEARN REPR
[2]  
Adebayo J, 2020, Arxiv, DOI arXiv:2011.05429
[3]  
Adebayo J, 2018, ADV NEUR IN, V31
[4]  
Alvarez-Melis D, 2018, Arxiv, DOI arXiv:1806.08049
[5]   Assessing the Trustworthiness of Saliency Maps for Localizing Abnormalities in Medical Imaging [J].
Arun, Nishanth ;
Gaw, Nathan ;
Singh, Praveer ;
Chang, Ken ;
Aggarwal, Mehak ;
Chen, Bryan ;
Hoebel, Katharina ;
Gupta, Sharut ;
Patel, Jay ;
Gidwani, Mishka ;
Adebayo, Julius ;
Li, Matthew D. ;
Kalpathy-Cramer, Jayashree .
RADIOLOGY-ARTIFICIAL INTELLIGENCE, 2021, 3 (06)
[6]  
Bethge M, 2019, Arxiv, DOI arXiv:1904.00760
[7]   Convolutional Dynamic Alignment Networks for Interpretable Classifications [J].
Boehle, Moritz ;
Fritz, Mario ;
Schiele, Bernt .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :10024-10033
[8]   Visual Explanations for the Detection of Diabetic Retinopathy from Retinal Fundus Images [J].
Boreiko, Valentyn ;
Ilanchezian, Indu ;
Seckin, Murat ;
Mueller, Ayhan Sarah ;
Koch, Lisa M. ;
Faber, Hanna ;
Berens, Philipp ;
Hein, Matthias .
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT II, 2022, 13432 :539-549
[9]  
Cohen JP, 2021, PR MACH LEARN RES, V143, P74
[10]  
Djoumessi K.R., 2023, arXiv