A non-convergent on-line training algorithm for neural networks

被引:0
|
作者
Utans, J [1 ]
机构
[1] London Business Sch, London NW1 4SA, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Stopped training is a method to avoid over-fitting of neural network models by preventing an iterative optimization method from reaching a local minimum of the objective function. It is motivated by the observation that over-fitting occurs gradually as training progresses. The stopping time is typically determined by monitoring the expected generalization performance of the model as approximated by the error on a validation set. In this paper we propose to use an analytic estimate for this purpose. However, these estimates require knowledge of the analytic form of the objective function used for training the network and are only applicable when the weights correspond to a local minimum of this objective function. For this reason, we propose the use of an auxiliary, regularized objective function. The algorithm is "self-contained" and does not require to split the data in a training and a separate validation set.
引用
收藏
页码:913 / 921
页数:9
相关论文
共 50 条
  • [21] On-line neural training algorithm with sliding mode control and adaptive learning rate
    Nied, A.
    Seleme, S. I., Jr.
    Parma, G. G.
    Menezes, B. R.
    NEUROCOMPUTING, 2007, 70 (16-18) : 2687 - 2691
  • [22] Simultaneously non-convergent frequencies of words in different expansions
    David Färm
    Monatshefte für Mathematik, 2011, 162 : 409 - 427
  • [23] A Deeper (Autoregressive) Approach to Non-Convergent Discourse Parsing
    Tulpan, Yoav
    Tsur, Oren
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 12883 - 12895
  • [24] An improved compound gradient vector based neural network on-line training algorithm
    Chen, ZP
    Dong, C
    Zhou, QQ
    Zhang, SJ
    DEVELOPMENTS IN APPLIED ARTIFICIAL INTELLIGENCE, 2003, 2718 : 316 - 325
  • [25] On-Line Extreme Learning Machine for Training Time-Varying Neural Networks
    Ye, Yibin
    Squartini, Stefano
    Piazza, Francesco
    BIO-INSPIRED COMPUTING AND APPLICATIONS, 2012, 6840 : 49 - 54
  • [26] Continual on-line training of neural networks with applications to electric machine fault diagnostics
    Tallam, RM
    Habetler, TG
    Harley, RG
    PESC 2001: 32ND ANNUAL POWER ELECTRONICS SPECIALISTS CONFERENCE, VOLS 1-4, CONFERENCE PROCEEDINGS, 2001, : 2224 - 2228
  • [27] Theoretical analysis of batch and on-line training for gradient descent learning in neural networks
    Nakama, Takehiko
    NEUROCOMPUTING, 2009, 73 (1-3) : 151 - 159
  • [28] On-line testing in digital neural networks
    Demidenko, S
    Piuri, V
    PROCEEDINGS OF THE FIFTH ASIAN TEST SYMPOSIUM (ATS '96), 1996, : 295 - 300
  • [29] A CASE OF NON-CONVERGENT DUAL CHANGES IN ASSIGNMENT PROBLEMS
    ARAOZ, J
    EDMONDS, J
    DISCRETE APPLIED MATHEMATICS, 1985, 11 (02) : 95 - 102
  • [30] Non-convergent boundaries and action ascription in multimodal interaction
    Gonzalez Temer, Veronica
    Ogden, Richard
    OPEN LINGUISTICS, 2021, 7 (01): : 685 - 706