A technical view on neural architecture search

被引:2
作者
Yi-Qi Hu
Yang Yu
机构
[1] Nanjing University,State Key Laboratory for Novel Software Technology
来源
International Journal of Machine Learning and Cybernetics | 2020年 / 11卷
关键词
Neural architecture search; AutoML; Deep learning; Machine learning;
D O I
暂无
中图分类号
学科分类号
摘要
Due to the discovery of innovative and practical neural architectures, deep learning has achieved bright successes in many fields, such as computer vision, natural language processing, recommendation systems, etc. To reach high performance, researchers have to adjust neural architectures and choose training tricks very carefully. The manual trial-and-error process for discovering the best neural network configuration consumes plenty of manpower. The neural architecture search (NAS) aims to alleviate this issue by automatically configuring neural networks. Recently, the rapid development of NAS has shown significant achievements. Novel neural network architectures that outperform the state-of-the-art handcrafted networks have been discovered in image classification benchmarks. In this paper, we survey NAS from a technical view. By summarizing the previous NAS approaches, we drew a picture of NAS for readers including problem definition, search approaches, progress towards practical applications and possible future directions. We hope that this paper can help beginners start their researches on NAS.
引用
收藏
页码:795 / 811
页数:16
相关论文
共 123 条
[1]  
Adankon MM(2009)Model selection for the LS-SVM. Application to handwriting recognition Pattern Recogn 42 3264-3270
[2]  
Cheriet M(1994)An evolutionary algorithm that constructs recurrent neural networks IEEE Trans Neural Netw 5 54-65
[3]  
Angeline PJ(1990)Evolution, learning, and culture: computational metaphors for adaptive algorithms Complex Syst 4 11-49
[4]  
Saunders GM(2000)Gradient-based optimization of hyperparameters Neural Comput 12 1889-1900
[5]  
Pollack JB(2013)Representation learning: a review and new perspectives IEEE Trans Pattern Anal Mach Intell 35 1798-1828
[6]  
Belew RK(2012)Random search for hyper-parameter optimization J Mach Learn Res 13 281-305
[7]  
Bengio Y(2003)Ranking learning algorithms: using IBL and meta-learning on accuracy and time results Mach Learn 50 251-277
[8]  
Bengio Y(1977)Pareto optimality in multiobjective problems Appl Math Optim 4 41-59
[9]  
Courville AC(2002)Model selection for small sample regression Mach Learn 48 9-23
[10]  
Vincent P(2019)Neural architecture search: a survey J Mach Learn Res 55 21-611