Exploring Neural Architecture Search Space via Deep Deterministic Sampling

被引:7
作者
Mills, Keith G. [1 ,2 ]
Salameh, Mohammad [2 ]
Niu, Di [1 ]
Han, Fred X. [2 ]
Rezaei, Seyed Saeed Changiz [2 ]
Yao, Hengshuai [2 ,3 ]
Lu, Wei [2 ]
Lian, Shuo [4 ]
Jui, Shangling [4 ]
机构
[1] Univ Alberta, Dept Elect & Comp Engn, Edmonton, AB T6G 2R3, Canada
[2] Huawei Technol Canada Co Ltd, Edmonton, AB T6G 2C8, Canada
[3] Univ Alberta, Dept Comp Sci, Edmonton, AB T6G 2R3, Canada
[4] Huawei Kirin Solut, Shanghai 201206, Peoples R China
关键词
Computer architecture; Optimization; Training; Reinforcement learning; Search problems; Stochastic processes; Graphics processing units; Neural architecture search; reinforcement learning; differentiable optimization;
D O I
10.1109/ACCESS.2021.3101975
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recent developments in Neural Architecture Search (NAS) resort to training the supernet of a predefined search space with weight sharing to speed up architecture evaluation. These include random search schemes, as well as various schemes based on optimization or reinforcement learning, in particular policy gradient, that aim to optimize a parametric architecture distribution and the shared model weights simultaneously. In this paper, we focus on efficiently exploring the important region of a neural architecture search space with reinforcement learning. We propose Deep Deterministic Architecture Sampling (DDAS) based on deep deterministic policy gradient and the actor-critic framework, to selectively sample important architectures in the supernet for training. Through balancing exploitation and exploration, DDAS is designed to combat the disadvantages of prior random supernet warm-up schemes and optimization schemes. Gradient-based NAS approaches require the execution of multiple short experiments in order to combat the random stochastic nature of gradient descent, while still only producing a single architecture. Contrary to this approach, DDAS employs a reinforcement learning-based agent and focuses on discovering a Pareto frontier containing many architectures over the course of a single experiment requiring 1 GPU day. Experimental results for CIFAR-10 and CIFAR-100 on the DARTS search space show that DDAS can depict in a single search, the accuracy-FLOPs (or model size) Pareto frontier, which outperforms random sampling and search. With a test accuracy of 97.27%, the best architecture found on CIFAR-10 outperforms the original second-order DARTS while using 600M fewer parameters. Additionally, DDAS finds an architecture capable of achieving 82.00% test accuracy on CIFAR-100 while using only 3.14M parameters and outperforming GDAS.
引用
收藏
页码:110962 / 110974
页数:13
相关论文
共 38 条
[1]  
[Anonymous], INT C LEARNING REPRE
[2]  
[Anonymous], ARXIV190111117
[3]  
[Anonymous], 2014, ICML ICML 14
[4]  
[Anonymous], 2017, P 2017 COMPUTER VISI
[5]  
Bender G, 2018, PR MACH LEARN RES, V80
[6]  
Cai H., 2020, P INT C LEARN REPR
[7]  
Cai Han, 2019, INT C LEARN REPR ICL
[8]   Progressive Differentiable Architecture Search: Bridging the Depth Gap between Search and Evaluation [J].
Chen, Xin ;
Xie, Lingxi ;
Wu, Jun ;
Tian, Qi .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :1294-1303
[9]  
Courbariaux M, 2015, ADV NEUR IN, V28
[10]   Searching for A Robust Neural Architecture in Four GPU Hours [J].
Dong, Xuanyi ;
Yang, Yi .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :1761-1770