Gradient Descent Effects on Differential Neural Architecture Search: A Survey

被引:25
作者
Santra, Santanu [1 ]
Hsieh, Jun-Wei [2 ]
Lin, Chi-Fang [1 ]
机构
[1] Yuan Ze Univ, Dept Comp Sci & Engn, Taoyuan 32003, Taiwan
[2] Natl Chiao Tung Univ, Coll Artificial Intelligence & Green Energy, Hsinchu 30010, Taiwan
关键词
Computer architecture; Search problems; Reinforcement learning; Training; Performance evaluation; Computational modeling; Task analysis; Gradient descent; neural architecture search; reinforcement learning; evolutionary algorithm; back-propagation;
D O I
10.1109/ACCESS.2021.3090918
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Gradient Descent, an effective way to search for the local minimum of a function, can minimize training and validation loss of neural architectures and also be incited in an appropriate order to decrease the searching cost of neural architecture search. In recent trends, the neural architecture search (NAS) is enormously used to construct an automatic architecture for a specific task. Mostly well-performed neural architecture search methods have adopted reinforcement learning, evolutionary algorithms, or gradient descent algorithms to find the best-performing candidate architecture. Among these methods, gradient descent-based architecture search approaches outperform all other methods in terms of efficiency, simplicity, computational cost, and validation error. In view of this, an in-depth survey is necessary to cover the usefulness of gradient descent method and how this can benefit neural architecture search. We begin our survey with basic concepts of neural architecture search, gradient descent, and their unique properties. Our survey then delves into the impact of gradient descent method on NAS and explores the effect of gradient descent in the search process to generate the candidate architecture. At the same time, our survey reviews mostly used gradient-based search approaches in NAS. Finally, we provide the current research challenges and open problems in the NAS-based approaches, which need to be addressed in future research.
引用
收藏
页码:89602 / 89618
页数:17
相关论文
共 86 条
[1]  
[Anonymous], 2017, INT C MACHINE LEARNI
[2]  
Ballester P, 2016, AAAI CONF ARTIF INTE, P1124
[3]  
Bender G, 2018, PR MACH LEARN RES, V80
[4]  
Brock A., 2018, ICLR
[5]  
Cai H., 2019, ICLR, P1
[6]  
Cai H, 2018, AAAI CONF ARTIF INTE, P2787
[7]  
Chang J., 2019, PROC ANN C NEURAL IN, P874
[8]  
Chen Xiangning, 2020, P INT C MACH LEARN, P1554
[9]   Progressive Differentiable Architecture Search: Bridging the Depth Gap between Search and Evaluation [J].
Chen, Xin ;
Xie, Lingxi ;
Wu, Jun ;
Tian, Qi .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :1294-1303
[10]   RENAS: Reinforced Evolutionary Neural Architecture Search [J].
Chen, Yukang ;
Meng, Gaofeng ;
Zhang, Qian ;
Xiang, Shiming ;
Huang, Chang ;
Mu, Lisen ;
Wang, Xinggang .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4782-4791