Stability analysis of gradient-based neural networks for optimization problems

被引:39
作者
Han, QM [1 ]
Liao, LZ
Qi, HD
Qi, LQ
机构
[1] Hong Kong Baptist Univ, Dept Math, Kowloon, Hong Kong, Peoples R China
[2] Nanjing Normal Univ, Sch Math & Comp Sci, Nanjing 210097, Peoples R China
[3] Univ New S Wales, Sch Math, Sydney, NSW 2052, Australia
[4] Hong Kong Polytech Univ, Dept Appl Math, Kowloon, Hong Kong, Peoples R China
基金
澳大利亚研究理事会;
关键词
gradient-based neural network; equilibrium point; equilibrium set; asymptotic stability; exponential stability;
D O I
10.1023/A:1011245911067
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
The paper introduces a new approach to analyze the stability of neural network models without using any Lyapunov function. With the new approach, we investigate the stability properties of the general gradient-based neural network model for optimization problems. Our discussion includes both isolated equilibrium points and connected equilibrium sets which could be unbounded. For a general optimization problem, if the objective function is bounded below and its gradient is Lipschitz continuous, we prove that (a) any trajectory of the gradient-based neural network converges to an equilibrium point, and (b) the Lyapunov stability is equivalent to the asymptotical stability in the gradient-based neural networks. For a convex optimization problem, under the same assumptions, we show that any trajectory of gradient-based neural networks will converge to an asymptotically stable equilibrium point of the neural networks. For a general nonlinear objective function, we propose a refined gradient-based neural network, whose trajectory with any arbitrary initial point will converge to an equilibrium point, which satisfies the second order necessary optimality conditions for optimization problems. Promising simulation results of a refined gradient-based neural network on some problems are also reported.
引用
收藏
页码:363 / 381
页数:19
相关论文
共 25 条
[1]   NEURAL NETWORK FOR QUADRATIC OPTIMIZATION WITH BOUND CONSTRAINTS [J].
BOUZERDOUM, A ;
PATTISON, TR .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1993, 4 (02) :293-304
[2]   Solving convex programming problems with equality constraints by neural networks [J].
Chen, YH ;
Fang, SC .
COMPUTERS & MATHEMATICS WITH APPLICATIONS, 1998, 36 (07) :41-68
[3]  
CHUA LO, 1984, IEEE T CIRCUITS SYST, V31, P182, DOI 10.1109/TCS.1984.1085482
[4]   NEURAL NETWORKS FOR COMPUTING EIGENVALUES AND EIGENVECTORS [J].
CICHOCKI, A ;
UNBEHAUEN, R .
BIOLOGICAL CYBERNETICS, 1992, 68 (02) :155-164
[5]   CONVEX PROGRAMMING IN HILBERT SPACE [J].
GOLDSTEIN, AA .
BULLETIN OF THE AMERICAN MATHEMATICAL SOCIETY, 1964, 70 (05) :709-&
[6]   A PROJECTION AND CONTRACTION METHOD FOR A CLASS OF LINEAR COMPLEMENTARITY-PROBLEMS AND ITS APPLICATION IN CONVEX QUADRATIC-PROGRAMMING [J].
HE, BS .
APPLIED MATHEMATICS AND OPTIMIZATION, 1992, 25 (03) :247-262
[7]  
He BS, 1997, APPL MATH OPT, V35, P69
[8]   A NEW METHOD FOR A CLASS OF LINEAR VARIATIONAL-INEQUALITIES [J].
HE, BS .
MATHEMATICAL PROGRAMMING, 1994, 66 (02) :137-144
[9]   SOLVING A CLASS OF LINEAR PROJECTION EQUATIONS [J].
HE, BS .
NUMERISCHE MATHEMATIK, 1994, 68 (01) :71-80
[10]  
HOPFIELD JJ, 1985, BIOL CYBERN, V52, P141