Extracting Explanations, Justification, and Uncertainty from Black-Box Deep Neural Networks

被引:0
作者
Ardis, Paul [1 ]
Flenner, Arjuna [2 ]
机构
[1] GE Aerosp Res, 1 Res Circle, Niskayuna, NY 12309 USA
[2] GE Aerosp, 3290 Patterson Ave SE, Grand Rapids, MI 49512 USA
来源
ASSURANCE AND SECURITY FOR AI-ENABLED SYSTEMS | 2024年 / 13054卷
关键词
D O I
10.1117/12.3012765
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Neural Networks (DNNs) do not inherently compute or exhibit empirically-justified task confidence. In mission critical applications, it is important to both understand associated DNN reasoning and its supporting evidence. In this paper, we propose a novel Bayesian approach to extract explanations, justifications, and uncertainty estimates from DNNs. Our approach is efficient both in terms of memory and computation, and can be applied to any black box DNN without any retraining, including applications to anomaly detection and out-of-distribution detection tasks. We validate our approach on the CIFAR-10 dataset, and show that it can significantly improve the interpretability and reliability of DNNs.
引用
收藏
页数:8
相关论文
共 50 条
  • [21] Practical Black-Box Attacks on Deep Neural Networks Using Efficient Query Mechanisms
    Bhagoji, Arjun Nitin
    He, Warren
    Li, Bo
    Song, Dawn
    COMPUTER VISION - ECCV 2018, PT XII, 2018, 11216 : 158 - 174
  • [22] A CMA-ES-Based Adversarial Attack on Black-Box Deep Neural Networks
    Kuang, Xiaohui
    Liu, Hongyi
    Wang, Ye
    Zhang, Qikun
    Zhang, Quanxin
    Zheng, Jun
    IEEE ACCESS, 2019, 7 : 172938 - 172947
  • [23] Black-box error diagnosis in Deep Neural Networks for computer vision: a survey of tools
    Piero Fraternali
    Federico Milani
    Rocio Nahime Torres
    Niccolò Zangrando
    Neural Computing and Applications, 2023, 35 : 3041 - 3062
  • [24] Generative causal explanations of black-box classifiers
    O'Shaughnessy, Matthew
    Canal, Gregory
    Connor, Marissa
    Davenport, Mark
    Rozell, Christopher
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [25] Learning Groupwise Explanations for Black-Box Models
    Gao, Jingyue
    Wang, Xiting
    Wang, Yasha
    Yan, Yulan
    Xie, Xing
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2396 - 2402
  • [26] Extracting Relational Explanations From Deep Neural Networks: A Survey From a Neural-Symbolic Perspective
    Townsend, Joe
    Chaton, Thomas
    Monteiro, Joao M.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (09) : 3456 - 3470
  • [27] EREBA: Black-box Energy Testing of Adaptive Neural Networks
    Haque, Mirazul
    Yadlapalli, Yaswanth
    Yang, Wei
    Liu, Cong
    2022 ACM/IEEE 44TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING (ICSE 2022), 2022, : 835 - 846
  • [28] Simple Black-Box Universal Adversarial Attacks on Deep Neural Networks for Medical Image Classification
    Koga, Kazuki
    Takemoto, Kazuhiro
    ALGORITHMS, 2022, 15 (05)
  • [29] DeepGD: A Multi-Objective Black-Box Test Selection Approach for Deep Neural Networks
    Aghababaeyan, Zohreh
    Abdellatif, Manel
    Dadkhah, Mahboubeh
    Briand, Lionel
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2024, 33 (06)
  • [30] NATTACK: Learning the Distributions of Adversarial Examples for an Improved Black-Box Attack on Deep Neural Networks
    Li, Yandong
    Li, Lijun
    Wang, Liqiang
    Zhang, Tong
    Gong, Boqing
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97