Intrinsic dimension of data representations in deep neural networks

被引:0
|
作者
Ansuini, Alessio [1 ]
Laio, Alessandro [1 ]
Macke, Jakob H. [2 ]
Zoccolan, Davide [1 ]
机构
[1] Scuola Int Super Studi Avanzati, Trieste, Italy
[2] Tech Univ Munich, Munich, Germany
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019) | 2019年 / 32卷
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks progressively transform their inputs across multiple processing layers. What are the geometrical properties of the representations learned by these networks? Here we study the intrinsic dimensionality (ID) of data-representations, i.e. the minimal number of parameters needed to describe a representation. We find that, in a trained network, the ID is orders of magnitude smaller than the number of units in each layer. Across layers, the ID first increases and then progressively decreases in the final layers. Remarkably, the ID of the last hidden layer predicts classification accuracy on the test set. These results can neither be found by linear dimensionality estimates (e.g., with principal component analysis), nor in representations that had been artificially linearized. They are neither found in untrained networks, nor in networks that are trained on randomized labels. This suggests that neural networks that can generalize are those that transform the data into low-dimensional, but not necessarily flat manifolds.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Neural networks for estimating intrinsic dimension
    Potapov, A
    Ali, MK
    PHYSICAL REVIEW E, 2002, 65 (04)
  • [2] The Tunnel Effect: Building Data Representations in Deep Neural Networks
    Masarczyk, Wojciech
    Ostaszewski, Mateusz
    Imani, Ehsan
    Pascanu, Razvan
    Milos, Piotr
    Trzcinski, Tomasz
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [3] Deep Convolutional Neural Network Compression based on the Intrinsic Dimension of the Training Data
    Hadi, Abir Mohammad
    Won, Kwanghee
    APPLIED COMPUTING REVIEW, 2024, 24 (01): : 14 - 23
  • [4] Data Dimension and Structure Effects in Predictive Performance of Deep Neural Networks
    Urda, Daniel
    Jerez, Jose M.
    Turias, Ignacio J.
    NEW TRENDS IN INTELLIGENT SOFTWARE METHODOLOGIES, TOOLS AND TECHNIQUES (SOMET_18), 2018, 303 : 361 - 372
  • [5] Deep Neural Networks for High Dimension, Low Sample Size Data
    Liu, Bo
    Wei, Ying
    Zhang, Yu
    Yang, Qiang
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2287 - 2293
  • [6] Investigating latent representations and generalization in deep neural networks for tabular data
    Couplet, Edouard
    Lambert, Pierre
    Verleysen, Michel
    Lee, John A.
    de Bodt, Cyril
    NEUROCOMPUTING, 2024, 597
  • [7] Intrinsic Dimension, Persistent Homology and Generalization in Neural Networks
    Birdal, Tolga
    Lou, Aaron
    Guibas, Leonidas
    Simsekli, Umut
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [8] Deep Networks as Paths on the Manifold of Neural Representations
    Lange, Richard D.
    Kwok, Devin
    Matelsky, Jordan
    Wang, Xinyue
    Rolnick, David
    Kording, Konrad P.
    TOPOLOGICAL, ALGEBRAIC AND GEOMETRIC LEARNING WORKSHOPS 2023, VOL 221, 2023, 221
  • [9] Deep Neural Networks for Learning Graph Representations
    Cao, Shaosheng
    Lu, Wei
    Xu, Qiongkai
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 1145 - 1152
  • [10] Exploring Internal Representations of Deep Neural Networks
    Despraz, Jeremie
    Gomez, Stephane
    Satizabal, Hector F.
    Pena-Reyes, Carlos Andres
    COMPUTATIONAL INTELLIGENCE, IJCCI 2017, 2019, 829 : 119 - 138