Topology-based representative datasets to reduce neural network training resources

被引:0
作者
Rocio Gonzalez-Diaz
Miguel A. Gutiérrez-Naranjo
Eduardo Paluzo-Hidalgo
机构
[1] Universidad de Sevilla,Dept. of Applied Mathematics I
[2] Universidad de Sevilla,Dept. of Computer Science and Artificial Intelligence
来源
Neural Computing and Applications | 2022年 / 34卷
关键词
Data reduction; Neural networks; Representative datasets; Computational topology;
D O I
暂无
中图分类号
学科分类号
摘要
One of the main drawbacks of the practical use of neural networks is the long time required in the training process. Such a training process consists of an iterative change of parameters trying to minimize a loss function. These changes are driven by a dataset, which can be seen as a set of labeled points in an n-dimensional space. In this paper, we explore the concept of a representative dataset which is a dataset smaller than the original one, satisfying a nearness condition independent of isometric transformations. Representativeness is measured using persistence diagrams (a computational topology tool) due to its computational efficiency. We theoretically prove that the accuracy of a perceptron evaluated on the original dataset coincides with the accuracy of the neural network evaluated on the representative dataset when the neural network architecture is a perceptron, the loss function is the mean squared error, and certain conditions on the representativeness of the dataset are imposed. These theoretical results accompanied by experimentation open a door to reducing the size of the dataset to gain time in the training process of any neural network.
引用
收藏
页码:14397 / 14413
页数:16
相关论文
共 16 条
  • [1] Wang Z(2020)Less is better: unweighted data subsampling via influence function Proc AAAI Conf Artif Intell 34 6340-6347
  • [2] Zhu H(2017)Computational aspects of the gromov-hausdorff distance and its application in non-rigid shape matching Discrete Comput Geom 57 854-880
  • [3] Dong Z(2014)Persistence stability for geometric complexes Geometriae Dedicata 173 193-214
  • [4] He X(2018)Representative datasets for neural networks Electron Notes Discrete Math 68 89-94
  • [5] Huang S-L(2019)Adversarial examples: attacks and defenses for deep learning IEEE Trans Neural Netw Learn Syst 30 2805-2824
  • [6] Schmiedl F(undefined)undefined undefined undefined undefined-undefined
  • [7] Chazal F(undefined)undefined undefined undefined undefined-undefined
  • [8] de Silva V(undefined)undefined undefined undefined undefined-undefined
  • [9] Oudot S(undefined)undefined undefined undefined undefined-undefined
  • [10] Gonzalez-Diaz R(undefined)undefined undefined undefined undefined-undefined