Layer factor analysis in convolutional neural networks for explainability

被引:1
作者
Lopez-Gonzalez, Clara I. [1 ]
Gomez-Silva, Maria J. [2 ]
Besada-Portas, Eva [2 ]
Pajares, Gonzalo [3 ]
机构
[1] Univ Complutense Madrid, Dept Software Engn & Artificial Intelligence, Madrid 28040, Spain
[2] Univ Complutense Madrid, Dept Comp Architecture & Automat, Madrid 28040, Spain
[3] Univ Complutense Madrid, Inst Knowledge Technol, Madrid 28040, Spain
关键词
Deep learning; Explainable artificial intelligence; Statistical modeling; Visual explanation; Feature learning; Attribution map; SCREE;
D O I
10.1016/j.asoc.2023.111094
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Explanatory methods that focus on the analysis of the features encoded by Convolutional Neural Networks (CNNs) are of great interest, since they help to understand the underlying process hidden behind the black box nature of these models. However, to explain the knowledge gathered in a given layer, they must decide which of the numerous filters to study, further assuming that each of them corresponds to a single feature. This, coupled with the redundancy of information, makes it difficult to ensure that the relevant characteristics are being analyzed. The above represents an important challenge and defines the aim and scope of our proposal. In this paper we present a novel method, named Explainable Layer Factor Analysis for CNNs (ELFACNNs), which models and describes with quality convolutional layers relying on factor analysis. Regarding contributions, ELFA obtains the essential underlying features, together with their correlation with the original filters, providing an accurate and well-founded summary. Through the factorial parameters we gain insights about the information learned, the connections between channels, and the redundancy of the layer, among others. To provide visual explanations in a similarly way to other methods, two additional proposals are made: a) Essential Feature Attribution Maps (EFAM) and b) intrinsic features inversion. The results prove the effectiveness of the developed general methods. They are evaluated in different CNNs (VGG-16, ResNet-50, and DeepLabv3+) on generic datasets (CIFAR-10, imagenette, and CamVid). We demonstrate that convolutional layers adequately fit a factorial model thanks to the new metrics presented for factor and fitting residuals (D1, D,, and Res, derive from covariance matrices). Moreover, knowledge about the deep image representations and the learning process is acquired, as well as reliable heat maps highlighting regions where essential features are located. This study effectively provides an explainable approach that can be applied to different CNNs and over different datasets.
引用
收藏
页数:17
相关论文
共 52 条
[1]  
Abhishek Kumar, 2022, Attribution-based xai methods in computer vision: A review
[2]  
Alber M, 2019, J MACH LEARN RES, V20
[3]  
[Anonymous], 2023, Keras applications: Modules
[4]  
[Anonymous], 2011, P 28 INT C MACH LEAR, DOI DOI 10.5555/3104482.3104528
[5]   On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation [J].
Bach, Sebastian ;
Binder, Alexander ;
Montavon, Gregoire ;
Klauschen, Frederick ;
Mueller, Klaus-Robert ;
Samek, Wojciech .
PLOS ONE, 2015, 10 (07)
[6]  
Barber D, 2012, Bayesian Reasoning and Machine Learning, DOI DOI 10.1017/CBO9780511804779
[7]   TESTS OF SIGNIFICANCE IN FACTOR ANALYSIS [J].
Bartlett, M. S. .
BRITISH JOURNAL OF PSYCHOLOGY-STATISTICAL SECTION, 1950, 3 :77-85
[8]   Semantic object classes in video: A high-definition ground truth database [J].
Brostow, Gabriel J. ;
Fauqueur, Julien ;
Cipolla, Roberto .
PATTERN RECOGNITION LETTERS, 2009, 30 (02) :88-97
[9]  
Carter S., 2019, Distill, V4, pe15, DOI DOI 10.23915/DISTILL.00015
[10]   SCREE TEST FOR NUMBER OF FACTORS [J].
CATTELL, RB .
MULTIVARIATE BEHAVIORAL RESEARCH, 1966, 1 (02) :245-276