Discretization-aware architecture search

被引:16
作者
Tian, Yunjie [1 ]
Liu, Chang [1 ]
Xie, Lingxi [2 ]
Jiao, Jianbin [1 ]
Ye, Qixiang [1 ]
机构
[1] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[2] Huawei Inc, Noahs Ark Lab, Beijing, Peoples R China
关键词
Neural architecture search; Weight-sharing; Discretization-aware; Imbalanced network configuration;
D O I
10.1016/j.patcog.2021.108186
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The search cost of neural architecture search (NAS) has been largely reduced by differentiable architecture search and weight-sharing methods. Such methods optimize a super-network with all possible edges and operations, and determine the optimal sub-network by discretization, i.e., pruning off operations/edges of small weights. However, the discretization process performed on either operations or edges incurs significant inaccuracy and thus the quality of the architecture is not guaranteed. In this paper, we propose discretization-aware architecture search (DA(2)S), and target at pushing the super-network towards the configuration of desired topology. DA(2)S is implemented with an entropy-based loss term, which can be regularized to differentiable architecture search in a plug-and-play fashion. The regularization is controlled by elaborated continuation functions, so that discretization is adaptive to the dynamic change of edges and operations. Experiments on standard image classification benchmarks demonstrate the effectiveness of our approach, in particular, under imbalanced network configurations that were not studied before. (C) 2021 Elsevier Ltd. All rights reserved.
引用
收藏
页数:12
相关论文
共 56 条
  • [1] [Anonymous], P 3 INT C LEARNING R
  • [2] [Anonymous], 2019, ICML
  • [3] Deep learning
    LeCun, Yann
    Bengio, Yoshua
    Hinton, Geoffrey
    [J]. NATURE, 2015, 521 (7553) : 436 - 444
  • [4] [Anonymous], 2017, COMMUN ACM, DOI DOI 10.1145/3065386
  • [5] [Anonymous], 2019, INT C MACH LEARN ICM
  • [6] Bender Gabriel, 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Proceedings, P14311, DOI 10.1109/CVPR42600.2020.01433
  • [7] Brock A., 2018, INT C LEARN REPR
  • [8] Cai H., 2019, INT C LEAR REPR ICLR
  • [9] Cai H, 2018, AAAI CONF ARTIF INTE, P2787
  • [10] Progressive DARTS: Bridging the Optimization Gap for NAS in the Wild
    Chen, Xin
    Xie, Lingxi
    Wu, Jun
    Tian, Qi
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2021, 129 (03) : 638 - 655