Intrinsic dimension of data representations in deep neural networks

被引:0
|
作者
Ansuini, Alessio [1 ]
Laio, Alessandro [1 ]
Macke, Jakob H. [2 ]
Zoccolan, Davide [1 ]
机构
[1] Scuola Int Super Studi Avanzati, Trieste, Italy
[2] Tech Univ Munich, Munich, Germany
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019) | 2019年 / 32卷
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks progressively transform their inputs across multiple processing layers. What are the geometrical properties of the representations learned by these networks? Here we study the intrinsic dimensionality (ID) of data-representations, i.e. the minimal number of parameters needed to describe a representation. We find that, in a trained network, the ID is orders of magnitude smaller than the number of units in each layer. Across layers, the ID first increases and then progressively decreases in the final layers. Remarkably, the ID of the last hidden layer predicts classification accuracy on the test set. These results can neither be found by linear dimensionality estimates (e.g., with principal component analysis), nor in representations that had been artificially linearized. They are neither found in untrained networks, nor in networks that are trained on randomized labels. This suggests that neural networks that can generalize are those that transform the data into low-dimensional, but not necessarily flat manifolds.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] Robustness of Sparsely Distributed Representations to Adversarial Attacks in Deep Neural Networks
    Sardar, Nida
    Khan, Sundas
    Hintze, Arend
    Mehra, Priyanka
    ENTROPY, 2023, 25 (06)
  • [42] Understanding Distributed Representations of Concepts in Deep Neural Networks without Supervision
    Chang, Wonjoon
    Kwon, Dahee
    Choi, Jaesik
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 10, 2024, : 11212 - 11220
  • [43] Deep convolutional neural networks for data delivery in vehicular networks
    Jiang, Hejun
    Tang, Xiaolan
    Jin, Kai
    Chen, Wenlong
    Pu, Juhua
    NEUROCOMPUTING, 2021, 432 (432) : 216 - 226
  • [44] Learning Deep Binaural Representations With Deep Convolutional Neural Networks for Spontaneous Speech Emotion Recognition
    Zhang, Shiqing
    Chen, Aihua
    Guo, Wenping
    Cui, Yueli
    Zhao, Xiaoming
    Liu, Limei
    IEEE ACCESS, 2020, 8 : 23496 - 23505
  • [45] Improving deep neural network design with new text data representations
    Prusa J.D.
    Khoshgoftaar T.M.
    Journal of Big Data, 4 (1)
  • [46] INTRINSIC DIMENSION OF GEOMETRIC DATA SETS
    Hanika, Tom
    Schneider, Friedrich Martin
    Stumme, Gerd
    TOHOKU MATHEMATICAL JOURNAL, 2022, 74 (01) : 23 - 52
  • [47] Understanding Deep Learning: Expected Spanning Dimension and Controlling the Flexibility of Neural Networks
    Berthiaume, David
    Paffenroth, Randy
    Guo, Lei
    FRONTIERS IN APPLIED MATHEMATICS AND STATISTICS, 2020, 6
  • [48] FDAM: full-dimension attention module for deep convolutional neural networks
    Cai, Silin
    Wang, Changping
    Ding, Jiajun
    Yu, Jun
    Fan, Jianping
    INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL, 2022, 11 (04) : 599 - 610
  • [49] FDAM: full-dimension attention module for deep convolutional neural networks
    Silin Cai
    Changping Wang
    Jiajun Ding
    Jun Yu
    Jianping Fan
    International Journal of Multimedia Information Retrieval, 2022, 11 : 599 - 610
  • [50] Dynamic Data Fault Localization for Deep Neural Networks
    Yin, Yining
    Feng, Yang
    Weng, Shihao
    Liu, Zixi
    Yao, Yuan
    Zhang, Yichi
    Zhao, Zhihong
    Chen, Zhenyu
    PROCEEDINGS OF THE 31ST ACM JOINT MEETING EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, ESEC/FSE 2023, 2023, : 1345 - 1357