Intrinsic dimension of data representations in deep neural networks

被引:0
|
作者
Ansuini, Alessio [1 ]
Laio, Alessandro [1 ]
Macke, Jakob H. [2 ]
Zoccolan, Davide [1 ]
机构
[1] Scuola Int Super Studi Avanzati, Trieste, Italy
[2] Tech Univ Munich, Munich, Germany
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019) | 2019年 / 32卷
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks progressively transform their inputs across multiple processing layers. What are the geometrical properties of the representations learned by these networks? Here we study the intrinsic dimensionality (ID) of data-representations, i.e. the minimal number of parameters needed to describe a representation. We find that, in a trained network, the ID is orders of magnitude smaller than the number of units in each layer. Across layers, the ID first increases and then progressively decreases in the final layers. Remarkably, the ID of the last hidden layer predicts classification accuracy on the test set. These results can neither be found by linear dimensionality estimates (e.g., with principal component analysis), nor in representations that had been artificially linearized. They are neither found in untrained networks, nor in networks that are trained on randomized labels. This suggests that neural networks that can generalize are those that transform the data into low-dimensional, but not necessarily flat manifolds.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] Comparing the Visual Representations and Performance of Humans and Deep Neural Networks
    Jacobs, Robert A.
    Bates, Christopher J.
    CURRENT DIRECTIONS IN PSYCHOLOGICAL SCIENCE, 2019, 28 (01) : 34 - 39
  • [22] Learning representations for the early detection of sepsis with deep neural networks
    Kam, Hye Jin
    Kim, Ha Young
    COMPUTERS IN BIOLOGY AND MEDICINE, 2017, 89 : 248 - 255
  • [23] Treeview and Disentangled Representations for Explaining Deep Neural Networks Decisions
    Sattigeri, Prasanna
    Ramamurthy, Karthikeyan Natesan
    Thiagarajan, Jayaraman J.
    Kailkhura, Bhavya
    2020 54TH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2020, : 284 - 288
  • [24] Relating Simple Sentence Representations in Deep Neural Networks and the Brain
    Jat, Sharmistha
    Tang, Hao
    Talukdar, Partha
    Mitchell, Tom
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 5137 - 5154
  • [25] Deep Learning Neural Networks and Bayesian Neural Networks in Data Analysis
    Chernoded, Andrey
    Dudko, Lev
    Myagkov, Igor
    Volkov, Petr
    XXIII INTERNATIONAL WORKSHOP HIGH ENERGY PHYSICS AND QUANTUM FIELD THEORY (QFTHEP 2017), 2017, 158
  • [26] Deep Neural Networks and Tabular Data: A Survey
    Borisov, Vadim
    Leemann, Tobias
    Sessler, Kathrin
    Haug, Johannes
    Pawelczyk, Martin
    Kasneci, Gjergji
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (06) : 7499 - 7519
  • [27] Deep Neural Networks Reveal a Gradient in the Complexity of Neural Representations across the Ventral Stream
    Guclu, Umut
    van Gerven, Marcel A. J.
    JOURNAL OF NEUROSCIENCE, 2015, 35 (27): : 10005 - 10014
  • [28] Weakly correlated synapses promote dimension reduction in deep neural networks
    Zhou, Jianwen
    Huang, Haiping
    PHYSICAL REVIEW E, 2021, 103 (01)
  • [29] Compression strategies and space-conscious representations for deep neural networks
    Marino, Giosue Cataldo
    Ghidoli, Gregorio
    Frasca, Marco
    Malchiodi, Dario
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 9835 - 9842
  • [30] Shared spatiotemporal category representations in biological and artificial deep neural networks
    Greene, Michelle R.
    Hansen, Bruce C.
    PLOS COMPUTATIONAL BIOLOGY, 2018, 14 (07)