Parameter convergence and learning curves for neural networks
被引:0
作者:
Fine, TL
论文数: 0引用数: 0
h-index: 0
机构:
Cornell Univ, Sch Elect Engn, Ithaca, NY 14853 USACornell Univ, Sch Elect Engn, Ithaca, NY 14853 USA
Fine, TL
[1
]
Mukherjee, S
论文数: 0引用数: 0
h-index: 0
机构:
Cornell Univ, Sch Elect Engn, Ithaca, NY 14853 USACornell Univ, Sch Elect Engn, Ithaca, NY 14853 USA
Mukherjee, S
[1
]
机构:
[1] Cornell Univ, Sch Elect Engn, Ithaca, NY 14853 USA
来源:
1998 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY - PROCEEDINGS
|
1998年
关键词:
D O I:
10.1109/ISIT.1998.708991
中图分类号:
TM [电工技术];
TN [电子技术、通信技术];
学科分类号:
0808 ;
0809 ;
摘要:
We revisit the off-studied asymptotic (in sample size) behavior of the parameter or weight estimate returned by any member of a large family of neural network training algorithms. By properly accounting for the characteristic property of neural networks that their empirical and generalization errors possess multiple minima, we establish conditions under which the parameter estimate converges strongly into the set of minima of the generalization error. These results are then used to derive learning curves for generalization and empirical errors that leads to bounds on rates of convergence.