Understanding neural network through neuron level visualization

被引:11
作者
Dou, Hui [1 ,2 ]
Shen, Furao [1 ,4 ]
Zhao, Jian [3 ]
Mu, Xinyu [1 ,4 ]
机构
[1] State Key Lab Novel Software Technol, Nanjing, Peoples R China
[2] Nanjing Univ, Dept Comp Sci & Technol, Nanjing 210023, Peoples R China
[3] Nanjing Univ, Sch Elect Sci & Engn, Nanjing 210023, Peoples R China
[4] Nanjing Univ, Sch Artificial Intelligence, Nanjing 210023, Peoples R China
基金
美国国家科学基金会;
关键词
Interpretability; Neural network; Visualization;
D O I
10.1016/j.neunet.2023.09.030
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neurons are the fundamental units of neural networks. In this paper, we propose a method for explaining neural networks by visualizing the learning process of neurons. For a trained neural network, the proposed method obtains the features learned by each neuron and displays the features in a human-understandable form. The features learned by different neurons are combined to analyze the working mechanism of different neural network models. The method is applicable to neural networks without requiring any changes to the architectures of the models. In this study, we apply the proposed method to both Fully Connected Networks (FCNs) and Convolutional Neural Networks (CNNs) trained using the backpropagation learning algorithm. We conduct experiments on models for image classification tasks to demonstrate the effectiveness of the method. Through these experiments, we gain insights into the working mechanisms of various neural network architectures and evaluate neural network interpretability from diverse perspectives.
引用
收藏
页码:484 / 495
页数:12
相关论文
共 29 条
[1]  
Nguyen A, 2016, ADV NEUR IN, V29
[2]   Network Dissection: Quantifying Interpretability of Deep Visual Representations [J].
Bau, David ;
Zhou, Bolei ;
Khosla, Aditya ;
Oliva, Aude ;
Torralba, Antonio .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :3319-3327
[3]  
Bodria F, 2021, Arxiv, DOI [arXiv:2102.13076, 10.48550/arXiv.2102.13076]
[4]  
Ghorbani A, 2019, AAAI CONF ARTIF INTE, P3681
[5]   Interpretable Learning for Self-Driving Cars by Visualizing Causal Attention [J].
Kim, Jinkyu ;
Canny, John .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :2961-2969
[6]   Explaining the black-box model: A survey of local interpretation methods for deep neural networks [J].
Liang, Yu ;
Li, Siguang ;
Yan, Chungang ;
Li, Maozhen ;
Jiang, Changjun .
NEUROCOMPUTING, 2021, 419 :168-182
[7]   The Mythos of Model Interpretability [J].
Lipton, Zachary C. .
COMMUNICATIONS OF THE ACM, 2018, 61 (10) :36-43
[8]   Ethical principles in machine learning and artificial intelligence: cases from the field and possible ways forward [J].
Lo Piano, Samuele .
HUMANITIES & SOCIAL SCIENCES COMMUNICATIONS, 2020, 7 (01)
[9]  
McGough Michael., 2018, Sacramento Bee
[10]  
Muhammad MB, 2020, IEEE IJCNN, DOI [10.51178/jsr.v1i1.14, 10.1109/ijcnn48605.2020.9206626]