From shallow feature learning to deep learning: Benefits from the width and depth of deep architectures

被引:66
作者
Zhong, Guoqiang [1 ]
Ling, Xiao [1 ]
Wang, Li-Na [1 ]
机构
[1] Ocean Univ China, Dept Comp Sci & Technol, Qingdao 266100, Peoples R China
基金
中国国家自然科学基金;
关键词
deep learning; feature learning; Image classification; object recognition; width and depth of deep architectures; NONLINEAR DIMENSIONALITY REDUCTION; NEURAL-NETWORKS; DISCRIMINANT-ANALYSIS; SPEECH RECOGNITION; COMPONENT ANALYSIS; REPRESENTATION; CLASSIFICATION; EIGENMAPS; MANIFOLDS;
D O I
10.1002/widm.1255
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Since Pearson developed principal component analysis (PCA) in 1901, feature learning (or called representation learning) has been studied for more than 100 years. During this period, many "shallow" feature learning methods have been proposed based on various learning criteria and techniques, until the popular deep learning research in recent years. In this advanced review, we describe the historical profile of the shallow feature learning research and introduce the important developments of the deep learning models. Particularly, we survey the deep architectures with benefits from the optimization of their width and depth, as these models have achieved new records in many applications, such as image classification and object detection. Finally, several interesting directions of deep learning are presented and briefly discussed. This article is categorized under: Technologies > Classification Technologies > Machine Learning Technologies > Computational Intelligence
引用
收藏
页数:14
相关论文
共 125 条
[1]  
[Anonymous], 2016, StackGAN: Text to photo-realistic image synthesis with stacked generative adversarial networks
[2]  
AULI M, 2013, P 2013 C EMP METH NA, P1044, DOI DOI 10.1057/AJP.2014.10
[3]   Generalized discriminant analysis using a kernel approach [J].
Baudat, G ;
Anouar, FE .
NEURAL COMPUTATION, 2000, 12 (10) :2385-2404
[4]   Laplacian eigenmaps for dimensionality reduction and data representation [J].
Belkin, M ;
Niyogi, P .
NEURAL COMPUTATION, 2003, 15 (06) :1373-1396
[5]  
Bengio P., 2006, Advances in Neural Information Processing Systems 19 (NIPS06), P153, DOI DOI 10.5555/2976456.2976476
[6]  
Bengio Y, 2001, ADV NEUR IN, V13, P932
[7]  
Bengio Y., 2003, NIPS, DOI DOI 10.5555/2981345.2981368
[8]   Representation Learning: A Review and New Perspectives [J].
Bengio, Yoshua ;
Courville, Aaron ;
Vincent, Pascal .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (08) :1798-1828
[9]  
Bengio Y, 2011, LECT NOTES ARTIF INT, V6925, P18, DOI 10.1007/978-3-642-24412-4_3
[10]   Learning Deep Architectures for AI [J].
Bengio, Yoshua .
FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2009, 2 (01) :1-127