Understanding the role of individual units in a deep neural network

被引:224
作者
Bau, David [1 ]
Zhu, Jun-Yan [1 ,2 ]
Strobelt, Hendrik [3 ]
Lapedriza, Agata [4 ,5 ]
Zhou, Bolei [6 ]
Torralba, Antonio [1 ]
机构
[1] MIT, Comp Sci & Artificial Intelligence Lab, 77 Massachusetts Ave, Cambridge, MA 02139 USA
[2] Adobe Inc, Adobe Res, San Jose, CA 95110 USA
[3] MIT, Int Business Machines IBM, Watson Artificial Intelligence Lab, Cambridge, MA 02142 USA
[4] MIT, Media Lab, Cambridge, MA 02139 USA
[5] Univ Oberta Catalunya, Estudis Informat Multimedia & Telecomunicacio, Barcelona 08018, Spain
[6] Chinese Univ Hong Kong, Dept Informat Engn, Shatin, Hong Kong, Peoples R China
关键词
machine learning; deep networks; computer vision;
D O I
10.1073/pnas.1907375117
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Deep neural networks excel at finding hierarchical representations that solve complex tasks over large datasets. How can we humans understand these learned representations? In this work, we present network dissection, an analytic framework to systematically identify the semantics of individual hidden units within image classification and image generation networks. First, we analyze a convolutional neural network (CNN) trained on scene classification and discover units that match a diverse set of object concepts. We find evidence that the network has learned many object classes that play crucial roles in classifying scene classes. Second, we use a similar analytic method to analyze a generative adversarial network (GAN) model trained to generate scenes. By analyzing changes made when small sets of units are activated or deactivated, we find that objects can be added and removed from the output scenes while adapting to the context. Finally, we apply our analytic framework to understanding adversarial attacks and to semantic image editing.
引用
收藏
页码:30071 / 30078
页数:8
相关论文
共 58 条
  • [1] Achille A, 2018, J MACH LEARN RES, V19
  • [2] [Anonymous], 2015, ARXIV150602078
  • [3] [Anonymous], Progressive growing of gans for improved quality, stability, and variation
  • [4] [Anonymous], 2017, IEEE I CONF COMP VIS, DOI DOI 10.1109/ICCV.2017.244
  • [5] [Anonymous], 2017, P 31 INT C NEURAL IN
  • [6] [Anonymous], 2016, PROC CVPR IEEE, DOI DOI 10.1109/CVPR.2016.90
  • [7] [Anonymous], 2016, ADV NEURAL INFORM PR
  • [8] On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation
    Bach, Sebastian
    Binder, Alexander
    Montavon, Gregoire
    Klauschen, Frederick
    Mueller, Klaus-Robert
    Samek, Wojciech
    [J]. PLOS ONE, 2015, 10 (07):
  • [9] Bau A., IDENTIFYING CONTROLL
  • [10] Bau D., 2020, CODE UNDERSTANDING R