Multi-Objective Neural Architecture Search by Learning Search Space Partitions

被引:0
作者
Zhao, Yiyang [1 ]
Wang, Linnan [2 ]
Guo, Tian [1 ]
机构
[1] Worcester Polytech Inst, Worcester, MA 01609 USA
[2] Brown Univ, Providence, RI USA
基金
美国国家科学基金会;
关键词
Neural Architecture Search; Monte Carlo Tree Search; AutoML; Deep Learning; EVOLUTIONARY ALGORITHMS; OPTIMIZATION; DIVERSITY; NETWORK;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deploying deep learning models requires taking into consideration neural network metrics such as model size, inference latency, and #FLOPs, aside from inference accuracy. This results in deep learning model designers leveraging multi -objective optimization to design effective deep neural networks in multiple criteria. However, applying multi -objective optimizations to neural architecture search (NAS) is nontrivial because NAS tasks usually have a huge search space, along with a non -negligible searching cost. This requires effective multi -objective search algorithms to alleviate the GPU costs. In this work, we implement a novel multi -objectives optimizer based on a recently proposed meta -algorithm called LaMOO Zhao et al. (2022) on NAS tasks. In a nutshell, LaMOO speedups the search process by learning a model from observed samples to partition the search space and then focusing on promising regions likely to contain a subset of the Pareto frontier. Using LaMOO , we observe an improvement of more than 200% sample efficiency compared to Bayesian optimization and evolutionary -based multi -objective optimizers on different NAS datasets. For example, when combined with LaMOO , qEHVI achieves a 225% improvement in sample efficiency compared to using qEHVI alone in NasBench201. For real -world tasks, LaMOO achieves 97.36% accuracy with only 1.62M #Params on CIFAR10 in only 600 search samples. On ImageNet, our large model reaches 80.4% top -1 accuracy with only 522M #FLOPs.
引用
收藏
页数:41
相关论文
共 94 条
  • [61] Tan MX, 2019, Arxiv, DOI arXiv:1907.09595
  • [62] Tan MX, 2019, PROC CVPR IEEE, P2815, DOI [arXiv:1807.11626, 10.1109/CVPR.2019.00293]
  • [63] FCOS: Fully Convolutional One-Stage Object Detection
    Tian, Zhi
    Shen, Chunhua
    Chen, Hao
    He, Tong
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 9626 - 9635
  • [64] Automatic Database Management System Tuning Through Large-scale Machine Learning
    Van Aken, Dana
    Pavlo, Andrew
    Gordon, Geoffrey J.
    Zhang, Bohan
    [J]. SIGMOD'17: PROCEEDINGS OF THE 2017 ACM INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA, 2017, : 1009 - 1024
  • [65] Van Veldhuizen D. A., 1998, P GEN PROGR 1998 C S, P221
  • [66] Wan A, 2020, PROC CVPR IEEE, P12962, DOI 10.1109/CVPR42600.2020.01298
  • [67] Searching the Deployable Convolution Neural Networks for GPUs
    Wang, Linnan
    Yu, Chenhan
    Salian, Satish
    Kierat, Slawomir
    Migacz, Szymon
    Florea, Alex Fit
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 12217 - 12226
  • [68] Wang LN, 2019, Arxiv, DOI arXiv:1903.11059
  • [69] Wang Linnan, 2019, arXiv
  • [70] Wang Lun, 2020, Advances in Neural Information Processing Systems, V33