A Learning Framework for Neural Networks Using Constrained Optimization Methods

被引:0
作者
Stavros J. Perantonis
Nikolaos Ampazis
Vassilis Virvilis
机构
[1] National Center for Scientific Research “Demokritos”,Institute of Informatics and Telecommunications
[2] National Center for Scientific Research “Demokritos”,Institute of Informatics and Telecommunications
来源
Annals of Operations Research | 2000年 / 99卷
关键词
neural networks; supervised learning; constrained optimization;
D O I
暂无
中图分类号
学科分类号
摘要
Conventional supervised learning in neural networks is carried out by performing unconstrained minimization of a suitably defined cost function. This approach has certain drawbacks, which can be overcome by incorporating additional knowledge in the training formalism. In this paper, two types of such additional knowledge are examined: Network specific knowledge (associated with the neural network irrespectively of the problem whose solution is sought) or problem specific knowledge (which helps to solve a specific learning task). A constrained optimization framework is introduced for incorporating these types of knowledge into the learning formalism. We present three examples of improvement in the learning behaviour of neural networks using additional knowledge in the context of our constrained optimization framework. The two network specific examples are designed to improve convergence and learning speed in the broad class of feedforward networks, while the third problem specific example is related to the efficient factorization of 2-D polynomials using suitably constructed sigma-pi networks.
引用
收藏
页码:385 / 401
页数:16
相关论文
共 26 条
  • [1] Ampazis N.(1999)Dynamics of multilayer networks in the vicinity of temporary minima Neural Networks 12 43-58
  • [2] Perantonis S.J.(1996)Does extra knowledge necessarily improve generalization? Neural Computation 8 202-214
  • [3] Taylor J.(1962)A steepest ascent method for solving optimum programming problems Journal App. Mech. 29 247-257
  • [4] Barber D.(1996)Learning with preknowledge: clustering with point and graph matching distance Neural Computation 8 787-804
  • [5] Saad D.(1988)Increased rates of convergence through learning rate adaptation Neural Networks 1 295-307
  • [6] Bryson A.E.(1992)Backpropagation learning for multilayer feedforward networks using the conjugate gradient method International Journal of Neural Systems 2 291-301
  • [7] Denham W.F.(1998)Constrained learning in neural networks: Application to stable factorization of 2-D polynomials Neural Proc. Lett. 7 5-14
  • [8] Gold S.(1995)An efficient learning algorithm with momentum acceleration Neural Networks 8 237-249
  • [9] Rangarajan A.(1993)A direct adaptive method for faster backpropagation learning: The RPROP algorithm Proceedings of the International Conference on Neural Networks 1 586-591
  • [10] Mjolsness E.(1996)Neural network training via an affine scaling quadratic optimization algorithm Neural Networks 9 475-481