Accelerating Evolutionary Neural Architecture Search via Multifidelity Evaluation

被引:13
作者
Yang, Shangshang [1 ]
Tian, Ye [2 ]
Xiang, Xiaoshu [2 ]
Peng, Shichen [1 ]
Zhang, Xingyi [1 ]
机构
[1] Anhui Univ, Sch Artificial Intelligence, Key Lab Intelligent Comp & Signal Proc, Minist Educ, Hefei 230039, Peoples R China
[2] Anhui Univ, Inst Phys Sci & Informat Technol, Key Lab Intelligent Comp & Signal Proc, Minist Educ, Hefei 230601, Peoples R China
基金
中国国家自然科学基金;
关键词
Computer architecture; Statistics; Evolutionary algorithms; Optimization; Costs; Convolutional neural networks; Graphics processing units; Convolutional neural networks (CNNs); evolutionary algorithm (EA); multifidelity evaluation; neural architecture search (NAS); MULTIOBJECTIVE OPTIMIZATION; GENETIC ALGORITHM; NETWORKS;
D O I
10.1109/TCDS.2022.3179482
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Evolutionary neural architecture search (ENAS) has recently received increasing attention by effectively finding high-quality neural architectures, which however consumes high computational cost by training the architecture encoded by each individual for complete epochs in individual evaluation. Numerous ENAS approaches have been developed to reduce the evaluation cost, but it is often difficult for most of these approaches to achieve high evaluation accuracy. To address this issue, in this article, we propose an accelerated ENAS via multifidelity evaluation termed MFENAS, where the individual evaluation cost is significantly reduced by training the architecture encoded by each individual for only a small number of epochs. The balance between evaluation cost and evaluation accuracy is well maintained by suggesting a multifidelity evaluation, which identifies the potentially good individuals that cannot survive from previous generations by integrating multiple evaluations under different numbers of training epochs. Besides, a population initialization strategy is devised to produce diverse neural architectures varying from ResNet-like architectures to Inception-like ones. As shown by experiments, the proposed MFENAS takes only 0.6 GPU days to find the best architecture holding a 2.39% test error rate, which is superior to most state-of-the-art neural architecture search approaches. And the architectures transferred to CIFAR-100 and ImageNet also exhibit competitive performance.
引用
收藏
页码:1778 / 1792
页数:15
相关论文
共 78 条
  • [1] Convolutional Neural Networks for Speech Recognition
    Abdel-Hamid, Ossama
    Mohamed, Abdel-Rahman
    Jiang, Hui
    Deng, Li
    Penn, Gerald
    Yu, Dong
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2014, 22 (10) : 1533 - 1545
  • [2] Automatic Design of Artificial Neural Networks for Gamma-Ray Detection
    Assuncao, Filipe
    Correia, Joao
    Conceicao, Ruben
    Martins Pimenta, Mario Joao
    Tome, Bernardo
    Lourenco, Nuno
    Machado, Penousal
    [J]. IEEE ACCESS, 2019, 7 : 110531 - 110540
  • [3] Fast DENSER: Efficient Deep NeuroEvolution
    Assuncao, Filipe
    Lourenco, Nuno
    Machado, Penousal
    Ribeiro, Bernardete
    [J]. GENETIC PROGRAMMING, EUROGP 2019, 2019, 11451 : 197 - 212
  • [4] Baker B, 2017, Arxiv, DOI [arXiv:1705.10823, 10.48550/arXiv.1705.10823]
  • [5] Baker B, 2017, Arxiv, DOI arXiv:1611.02167
  • [6] A simulated annealing-based multiobjective optimization algorithm: AMOSA
    Bandyopadhyay, Sanghamitra
    Saha, Sriparna
    Maulik, Ujjwal
    Deb, Kalyanmoy
    [J]. IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2008, 12 (03) : 269 - 283
  • [7] Cai H., 2018, PROC AAAI, P340
  • [8] RENAS: Reinforced Evolutionary Neural Architecture Search
    Chen, Yukang
    Meng, Gaofeng
    Zhang, Qian
    Xiang, Shiming
    Huang, Chang
    Mu, Lisen
    Wang, Xinggang
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4782 - 4791
  • [9] Coello CA, 2007, EVOLUTIONARY ALGORIT, V5
  • [10] Evolutionary multi-objective optimization: A historical view of the field
    Coello Coello, Carlos A.
    [J]. IEEE COMPUTATIONAL INTELLIGENCE MAGAZINE, 2006, 1 (01) : 28 - 36