OCIE: Augmenting model interpretability via Deconfounded Explanation-Guided Learning

被引:0
作者
Dong, Liang
Chen, Leiyang
Zheng, Chengliang
Fu, Zhongwang
Zukaib, Umer
Cui, Xiaohui [1 ]
Shen, Zhidong
机构
[1] Wuhan Univ, Sch Cyber Sci & Engn, Wuhan 430000, Peoples R China
关键词
Explanation-guided learning; Causal inference; Explainable Artificial Intelligence; Unsupervised object detection; Image classification; ATTENTION; TRANSFORMER;
D O I
10.1016/j.knosys.2024.112390
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) often encounter significant challenges related to opacity, inherent biases, and shortcut learning, which undermine their practical reliability. In this study, we address these issues by constructing a causal graph to model the unbiased learning process of DNNs. This model reveals that recurrent background information in training samples acts as a confounder, leading to spurious correlations between model inputs and outputs, causing biased predictions. To mitigate these problems and promote unbiased feature learning, we propose the Object-guided Consistency Interpretation Enhancement (OCIE) methodology. OCIE enhances DNN interpretability by integrating explicit objects and explanations into the model's learning process. Initially, OCIE employs a graph-based algorithm to identify explicit objects within self-supervised vision transformer-learned features. Subsequently, it constructs class prototypes to eliminate invalid detected objects. Finally, OCIE aligns explanations with explicit objects, directing the model's attention towards the most distinctive classification features rather than irrelevant backgrounds. Extensive experiments on different image classification datasets, including general (ImageNet), fine-grained (Stanford Cars and CUB-200), and medical (HAM) datasets, using two prevailing network architectures, demonstrate that OCIE significantly enhances explanation consistency across all datasets. Furthermore, OCIE proves particularly advantageous for fine-grained classification, especially in few-shot scenarios, by improving both interpretability and classification performance. Additionally, our findings highlight the impact of centralized explanations on the sufficiency of model decisions, suggesting that focusing explanations on explicit objects improves the reliability of DNN predictions. Our code is available at: https://github.com/DLAIResearch/OCIE.
引用
收藏
页数:24
相关论文
共 79 条
  • [1] Abnar S, 2020, Arxiv, DOI arXiv:2005.00928
  • [2] Adebayo J, 2018, ADV NEUR IN, V31
  • [3] Augustin Maximilian, 2022, ADV NEUR IN
  • [4] The Fast Bilateral Solver
    Barron, Jonathan T.
    Poole, Ben
    [J]. COMPUTER VISION - ECCV 2016, PT III, 2016, 9907 : 617 - 632
  • [5] A survey on XAI and natural language explanations
    Cambria, Erik
    Malandri, Lorenzo
    Mercorio, Fabio
    Mezzanzanica, Mario
    Nobani, Navid
    [J]. INFORMATION PROCESSING & MANAGEMENT, 2023, 60 (01)
  • [6] Cao S., 2024, Adv. Neural Inf. Process. Syst., V36, DOI [10.48550/arXiv.2402.03311, DOI 10.48550/ARXIV.2402.03311]
  • [7] Emerging Properties in Self-Supervised Vision Transformers
    Caron, Mathilde
    Touvron, Hugo
    Misra, Ishan
    Jegou, Herve
    Mairal, Julien
    Bojanowski, Piotr
    Joulin, Armand
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 9630 - 9640
  • [8] Emotion Recognition Using Three-Dimensional Feature and Convolutional Neural Network from Multichannel EEG Signals
    Chao, Hao
    Dong, Liang
    [J]. IEEE SENSORS JOURNAL, 2021, 21 (02) : 2024 - 2034
  • [9] Transformer Interpretability Beyond Attention Visualization
    Chefer, Hila
    Gur, Shir
    Wolf, Lior
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 782 - 791
  • [10] Milestones in Autonomous Driving and Intelligent Vehicles: Survey of Surveys
    Chen, Long
    Li, Yuchen
    Huang, Chao
    Li, Bai
    Xing, Yang
    Tian, Daxin
    Li, Li
    Hu, Zhongxu
    Na, Xiaoxiang
    Li, Zixuan
    Teng, Siyu
    Lv, Chen
    Wang, Jinjun
    Cao, Dongpu
    Zheng, Nanning
    Wang, Fei-Yue
    [J]. IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2023, 8 (02): : 1046 - 1056