On the Size of Convolutional Neural Networks and Generalization Performance

被引:0
作者
Kabkab, Maya [1 ,2 ]
Hand, Emily [1 ]
Chellappa, Rama [1 ,2 ]
机构
[1] Univ Maryland, Ctr Automat Res, UMIACS, College Pk, MD 20742 USA
[2] Univ Maryland, Dept Elect & Comp Engn, College Pk, MD 20742 USA
来源
2016 23RD INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR) | 2016年
关键词
COMPLEXITY;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While Convolutional Neural Networks (CNNs) have recently achieved impressive results on many classification tasks, it is still unclear why they perform so well and how to properly design them. In this work, we investigate the effect of the convolutional depth of a CNN on its generalization performance for binary classification problems. We prove a sufficient condition-polynomial in the depth of the CNN-on the training database size to guarantee such performance. We empirically test our theory on the problem of gender classification and explore the effect of varying the CNN depth, as well as the training distribution and set size.
引用
收藏
页码:3572 / 3577
页数:6
相关论文
共 25 条
[1]  
[Anonymous], 2008, ECCV
[2]  
[Anonymous], 2014, WORKSH INT C LEARN R
[3]  
[Anonymous], 2015, P IEEE C COMPUTER VI
[4]  
[Anonymous], 2014, EUROPEAN C COMPUTER
[5]  
[Anonymous], 2000, NATURE STAT LEARNING, DOI DOI 10.1007/978-1-4757-3264-1
[6]  
[Anonymous], IEEE ICCV WORKSH
[7]  
[Anonymous], 2014, NIPS
[8]  
[Anonymous], IEEE ICCV
[9]  
Bartlett P. L., 2003, HDB BRAIN THEORY NEU, P1188
[10]   The sample complexity of pattern classification with neural networks: The size of the weights is more important than the size of the network [J].
Bartlett, PL .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1998, 44 (02) :525-536