Solving the linear interval tolerance problem for weight initialization of neural networks

被引:29
作者
Adam, S. P. [1 ,2 ]
Karras, D. A. [3 ]
Magoulas, G. D. [4 ]
Vrahatis, M. N. [1 ]
机构
[1] Univ Patras, Dept Math, Computat Intelligence Lab, GR-26110 Patras, Greece
[2] Technol Educ Inst Epirus, Dept Comp Engn, Arta 47100, Greece
[3] Technol Educ Inst Sterea Hellas, Dept Automat, Psahna 34400, Evia, Greece
[4] Univ London, Birkbeck Coll, Dept Comp Sci & Informat Syst, London WC1E 7HX, England
关键词
Neural networks; Weight initialization; Interval analysis; Linear interval tolerance problem; FEEDFORWARD NETWORKS; STATISTICAL TESTS; TRAINING SPEED; HIGH-DIMENSION; BACKPROPAGATION; ALGORITHM; INTELLIGENCE;
D O I
10.1016/j.neunet.2014.02.006
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Determining good initial conditions for an algorithm used to train a neural network is considered a parameter estimation problem dealing with uncertainty about the initial weights. Interval analysis approaches model uncertainty in parameter estimation problems using intervals and formulating tolerance problems. Solving a tolerance problem is defining lower and upper bounds of the intervals so that the system functionality is guaranteed within predefined limits. The aim of this paper is to show how the problem of determining the initial weight intervals of a neural network can be defined in terms of solving a linear interval tolerance problem. The proposed linear interval tolerance approach copes with uncertainty about the initial weights without any previous knowledge or specific assumptions on the input data as required by approaches such as fuzzy sets or rough sets. The proposed method is tested on a number of well known benchmarks for neural networks trained with the back-propagation family of algorithms. Its efficiency is evaluated with regards to standard performance measures and the results obtained are compared against results of a number of well known and established initialization methods. These results provide credible evidence that the proposed method outperforms classical weight initialization methods. (C) 2014 Elsevier Ltd. All rights reserved.
引用
收藏
页码:17 / 37
页数:21
相关论文
共 50 条
  • [31] Improving learning in neural networks through weight initializations
    Mittal, Apeksha
    Singh, Amit Prakash
    Chandra, Pravin
    JOURNAL OF INFORMATION & OPTIMIZATION SCIENCES, 2021, 42 (05) : 951 - 971
  • [32] A comparison of linear and morphological shared-weight neural networks
    Won, Y
    Gader, PD
    NONLINEAR IMAGE PROCESSING VII, 1996, 2662 : 81 - 92
  • [33] Solving Linear Equations With Separable Problem Data Over Directed Networks
    Srivastava, Priyank
    Cortes, Jorge
    IEEE CONTROL SYSTEMS LETTERS, 2022, 6 : 596 - 601
  • [34] Design of a kind of nonlinear neural networks for solving the inverse optimal value problem with convex constraints
    Wu, Huaiqin
    Wang, Kewang
    Guo, Qiangqiang
    Xu, Guohua
    Li, Ning
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2014, 5 (01) : 85 - 92
  • [35] A novel weight initialization method for the random neural network
    Timotheou, Stelios
    NEUROCOMPUTING, 2009, 73 (1-3) : 160 - 168
  • [36] Solving an Optimal Control Problem of Cancer Treatment by Artificial Neural Networks
    Heydarpour, F.
    Abbasi, E.
    Ebadi, M. J.
    Karbassi, S. M.
    INTERNATIONAL JOURNAL OF INTERACTIVE MULTIMEDIA AND ARTIFICIAL INTELLIGENCE, 2020, 6 (04): : 18 - 25
  • [37] Solving maximum fuzzy clique problem with neural networks and its applications
    Bhattacharyya M.
    Bandyopadhyay S.
    Memetic Computing, 2009, 1 (4) : 281 - 290
  • [38] PROSPECTS FOR THE USE OF ARTIFICIAL NEURAL NETWORKS FOR PROBLEM SOLVING IN CLINICAL TRANSPLANTATION
    Kurabekova, R. M.
    Belchenkov, A. A.
    Shevchenko, O. P.
    VESTNIK TRANSPLANTOLOGII I ISKUSSTVENNYH ORGANOV, 2021, 23 (02): : 177 - 182
  • [39] AN INITIALIZATION METHOD FOR FEEDFORWARD ARTIFICIAL NEURAL NETWORKS USING POLYNOMIAL BASES
    Varnava, Thanasis M.
    Meade, Andrew J., Jr.
    ADVANCES IN DATA SCIENCE AND ADAPTIVE ANALYSIS, 2011, 3 (03) : 385 - 400
  • [40] Solving the N-bit parity problem using neural networks
    Hohil, ME
    Liu, DR
    Smith, SH
    NEURAL NETWORKS, 1999, 12 (09) : 1321 - 1323