Investigating over-parameterized randomized graph networks

被引:0
作者
Donghi, Giovanni [1 ]
Pasa, Luca [1 ]
Oneto, Luca [2 ]
Gallicchio, Claudio [3 ]
Micheli, Alessio [3 ]
Anguita, Davide [2 ]
Sperduti, Alessandro [1 ]
Navarin, Nicolo [1 ]
机构
[1] Univ Padua, Via Trieste 63, I-35121 Padua, Italy
[2] Univ Genoa, Via Opera Pia 11a, I-16145 Genoa, Italy
[3] Univ Pisa, Largo B Pontecorvo 3, I-56127 Pisa, Italy
关键词
Graph neural networks; Deep randomized neural networks; Algorithmic stability; Over-parameterization; STABILITY;
D O I
10.1016/j.neucom.2024.128281
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we investigate neural models based on graph random features for classification tasks. First, we aim to understand when over parameterization, namely generating more features than the ones necessary to interpolate, may be beneficial for the generalization abilities of the resulting models. We employ two measures: one from the algorithmic stability framework and another one based on information theory. We provide empirical evidence from several commonly adopted graph datasets showing that the considered measures, even without considering task labels, can be effective for this purpose. Additionally, we investigate whether these measures can aid in the process of hyperparameters selection. The results of our empirical analysis show that the considered measures have good correlations with the estimated generalization performance of the models with different hyperparameter configurations. Moreover, they can be used to identify good hyperparameters, achieving results comparable to the ones obtained with a classic grid search.
引用
收藏
页数:12
相关论文
共 45 条
  • [1] Achiam OJ, 2023, Arxiv, DOI [arXiv:2303.08774, 10.48550/arXiv.2303.08774]
  • [2] Pyramidal Reservoir Graph Neural Network
    Bianchi, F. M.
    Gallicchio, Claudio
    Micheli, Alessio
    [J]. NEUROCOMPUTING, 2022, 470 : 389 - 404
  • [3] Protein function prediction via graph kernels
    Borgwardt, KM
    Ong, CS
    Schönauer, S
    Vishwanathan, SVN
    Smola, AJ
    Kriegel, HP
    [J]. BIOINFORMATICS, 2005, 21 : I47 - I56
  • [4] Stability and generalization
    Bousquet, O
    Elisseeff, A
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2002, 2 (03) : 499 - 526
  • [5] Optimizing Reservoir Computers for Signal Classification
    Carroll, Thomas L.
    [J]. FRONTIERS IN PHYSIOLOGY, 2021, 12
  • [6] Chen ZJ, 2021, Arxiv, DOI arXiv:2110.11477
  • [7] ROCKET: exceptionally fast and accurate time series classification using random convolutional kernels
    Dempster, Angus
    Petitjean, Francois
    Webb, Geoffrey, I
    [J]. DATA MINING AND KNOWLEDGE DISCOVERY, 2020, 34 (05) : 1454 - 1495
  • [8] DISTRIBUTION-FREE INEQUALITIES FOR THE DELETED AND HOLDOUT ERROR ESTIMATES
    DEVROYE, LP
    WAGNER, TJ
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 1979, 25 (02) : 202 - 207
  • [9] Elisseeff A, 2005, J MACH LEARN RES, V6, P55
  • [10] Errica Federico, 2019, P 8 INT C LEARN REPR