Prioritized Architecture Sampling with Monto-Carlo Tree Search

被引:35
作者
Su, Xiu [1 ]
Huang, Tao [2 ]
Li, Yanxi [1 ]
You, Shan [2 ,3 ]
Wang, Fei [2 ]
Qian, Chen [2 ]
Zhang, Changshui [3 ]
Xu, Chang [1 ]
机构
[1] Univ Sydney, Fac Engn, Sch Comp Sci, Sydney, NSW, Australia
[2] SenseTime Res, Hong Kong, Peoples R China
[3] Tsinghua Univ, Dept Automat, Beijing, Peoples R China
来源
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021 | 2021年
基金
澳大利亚研究理事会;
关键词
D O I
10.1109/CVPR46437.2021.01082
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
One-shot neural architecture search (NAS) methods significantly reduce the search cost by considering the whole search space as one network, which only needs to be trained once. However, current methods select each operation independently without considering previous layers. Besides, the historical information obtained with huge computation costs is usually used only once and then discarded. In this paper, we introduce a sampling strategy based on Monte Carlo tree search (MCTS) with the search space modeled as a Monte Carlo tree (MCT), which captures the dependency among layers. Furthermore, intermediate results are stored in the MCT for future decisions and a better exploration-exploitation balance. Concretely, MCT is updated using the training loss as a reward to the architecture performance; for accurately evaluating the numerous nodes, we propose node communication and hierarchical node selection methods in the training and search stages, respectively, making better uses of the operation rewards and hierarchical information. Moreover, for a fair comparison of different NAS methods, we construct an opensource NAS benchmark of a macro search space evaluated on CIFAR-10, namely NAS-Bench-Macro. Extensive experiments on NAS-Bench-Macro and ImageNet demonstrate that our method significantly improves search efficiency and performance. For example, by only searching 20 architectures, our obtained architecture achieves 78.0% top-1 accuracy with 442M FLOPs on ImageNet.
引用
收藏
页码:10963 / 10972
页数:10
相关论文
共 37 条
[1]  
Cai H., 2018, ARXIV181200332
[2]  
Chu X., 2019, ARXIV190701845
[3]  
Chu Xiangxiang, 2019, ARXIV190806022
[4]   A fast and elitist multiobjective genetic algorithm: NSGA-II [J].
Deb, K ;
Pratap, A ;
Agarwal, S ;
Meyarivan, T .
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2002, 6 (02) :182-197
[5]  
Du S., 2020, ADV NEURAL INFORM PR, V33
[6]  
Fang Muyuan, 2019, ARXIV191211191
[7]   Powering One-Shot Topological NAS with Stabilized Share-Parameter Proxy [J].
Guo, Ronghao ;
Lin, Chen ;
Li, Chuming ;
Tian, Keyu ;
Sun, Ming ;
Sheng, Lu ;
Yan, Junjie .
COMPUTER VISION - ECCV 2020, PT XIV, 2020, 12359 :625-641
[8]   Stretchable Textile Rechargeable Zn Batteries Enabled by a Wax Dyeing Method [J].
Guo, Zi Hao ;
Liu, Mengmeng ;
Cong, Zifeng ;
Guo, Wenbin ;
Zhang, Panpan ;
Hu, Weiguo ;
Pu, Xiong .
ADVANCED MATERIALS TECHNOLOGIES, 2020, 5 (11)
[9]  
Han K, 2021, ABS210300112 ARXIV
[10]   GhostNet: More Features from Cheap Operations [J].
Han, Kai ;
Wang, Yunhe ;
Tian, Qi ;
Guo, Jianyuan ;
Xu, Chunjing ;
Xu, Chang .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :1577-1586