Prior knowledge and preferential structures in gradient descent learning algorithms

被引:12
作者
Mahony, RE
Williamson, RC
机构
[1] Australian Natl Univ, Dept Engn, Canberra, ACT 0200, Australia
[2] Australian Natl Univ, Dept Telecommun Engn, Res Sch Informat Sci & Engn, Canberra, ACT 0200, Australia
关键词
gradient descent; exponentiated gradient algorithm; natural gradient; link-functions; Riemannian metric;
D O I
10.1162/153244301753683735
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A family of gradient descent algorithms for learning linear functions in an online setting is considered. The family includes the classical LMS algorithm as well as new variants such as the Exponentiated Gradient (EG) algorithm due to Kivinen and Warmuth. The algorithms are based on prior distributions defined on the weight space. Techniques from differential geometry are used to develop the algorithms as gradient descent iterations with respect to the natural gradient in the Riemannian structure induced by the prior distribution. The proposed framework subsumes the notion of "link-functions".
引用
收藏
页码:311 / 355
页数:45
相关论文
共 40 条
  • [1] Abbot EdwinA., 1992, Flatland: A Romance in Many Dimensions
  • [2] AKAIKE H, 1980, J ROY STAT SOC B MET, V42, P46
  • [3] Natural gradient works efficiently in learning
    Amari, S
    [J]. NEURAL COMPUTATION, 1998, 10 (02) : 251 - 276
  • [4] AMARI S, 1997, ADV NEURAL INFORMATI, V9
  • [5] Amari S., 1985, DIFFERENTIAL GEOMETR
  • [6] [Anonymous], 1999, FEEDFORWARD NEURAL N
  • [7] [Anonymous], 1997, RIEMANNIAN MANIFOLDS
  • [8] [Anonymous], UNSUPERVISED ADAPT 1
  • [9] BARNDORFFNIELSE.OE, 1988, PARAMETRIC STAT MODE
  • [10] Boothby W.M., 1986, INTRO DIFFERENTIABLE