To go deep or wide in learning?

被引:0
作者
Pandey, Gaurav [1 ]
Dukkipati, Ambedkar [1 ]
机构
[1] Indian Inst Sci, Dept Comp Sci & Automat, Bangalore 560012, Karnataka, India
来源
ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 33 | 2014年 / 33卷
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To achieve acceptable performance for AI tasks, one can either use sophisticated feature extraction methods as the first layer in a two-layered supervised learning model, or learn the features directly using a deep (multilayered) model. While the first approach is very problem-specific, the second approach has computational overheads in learning multiple layers and fine-tuning of the model. In this paper, we propose an approach called wide learning based on arc-cosine kernels, that learns a single layer of infinite width. We propose exact and inexact learning strategies for wide learning and show that wide learning with single layer outperforms single layer as well as deep architectures of finite width for some benchmark datasets.
引用
收藏
页码:724 / 732
页数:9
相关论文
共 23 条
  • [1] [Anonymous], 2012, NIPS
  • [2] [Anonymous], 2005, PROC CVPR IEEE
  • [3] [Anonymous], 2010, ADV NEUR INF PROC SY
  • [4] [Anonymous], 2007, IEEE INT C ICML
  • [5] [Anonymous], P 21 INT C MACH LEAR
  • [6] [Anonymous], 2013, PMLR, DOI DOI 10.5555/3042817.3043055
  • [7] Bengio Y., 2006, Advances in Neural Information Processing Systems, V19, DOI DOI 10.7551/MITPRESS/7503.003.0024
  • [8] Learning Deep Architectures for AI
    Bengio, Yoshua
    [J]. FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2009, 2 (01): : 1 - 127
  • [9] Bergstra J., 2011, ADV NEURAL INF PROCE
  • [10] Large-Margin Classification in Infinite Neural Networks
    Cho, Youngmin
    Saul, Lawrence K.
    [J]. NEURAL COMPUTATION, 2010, 22 (10) : 2678 - 2697