Efficient Self-learning Evolutionary Neural Architecture Search

被引:7
作者
Qiu, Zhengzhong [1 ]
Bi, Wei [2 ]
Xu, Dong [3 ,4 ]
Guo, Hua [2 ]
Ge, Hongwei [5 ]
Liang, Yanchun [6 ]
Lee, Heow Pueh [7 ]
Wu, Chunguo [1 ]
机构
[1] Jilin Univ, Minist Educ, Coll Comp Sci & Technol, Key Lab Symbol Computat & Knowledge Engn, Changchun 130012, Jilin, Peoples R China
[2] YGSOFT INC, Zhuhai 519085, Guangdong, Peoples R China
[3] Univ Missouri, Dept Elect Engn & Comp Sci, Columbia, MO 65211 USA
[4] Univ Missouri, Christopher S Bond Life Sci Ctr, Columbia, MO 65211 USA
[5] Dalian Univ Technol, Coll Comp Sci & Technol, Dalian 116081, Liaoning, Peoples R China
[6] Zhuhai Coll Sci & Technol, Sch Comp Sci, Zhuhai 519041, Guangdong, Peoples R China
[7] Natl Univ Singapore, Dept Mech Engn, 9 Engn Dr 1, Singapore 117575, Singapore
基金
中国国家自然科学基金;
关键词
Evolutionary algorithm; Neural architecture search; Probability distribution; Model size control; NETWORKS;
D O I
10.1016/j.asoc.2023.110671
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The evolutionary algorithm has become a major method for neural architecture search recently. However, the fixed probability distribution employed by the traditional evolutionary algorithm may lead to structural complexity and redundancy due to its inability to control the size of individual architectures, and it cannot learn from empirical information gathered during the search process to guide the subsequent search more effectively and efficiently. Moreover, evaluating the performance of all the searched architectures requires significant computing resources and time overhead. To overcome these challenges, we present the Efficient Self-learning Evolutionary Neural Architecture Search (ESE-NAS) method. Firstly, we propose an Adaptive Learning Strategy for Mutation Sampling, composed of a Model Size Control module and a Credit Assignment method for Mutation Candidates, to guide the search process by learning from the model size information and evaluation results of the architectures and adjusting the probability distributions for evolution sampling accordingly. Additionally, we developed a neural architecture performance predictor to further improve the efficiency of NAS. Experiments on CIFAR-10 and CIFAR-100 datasets show that ESE-NAS significantly brings forward the first hitting time of the optimal architectures and reaches a competitive performance level with classic manual-designed and NAS models while maintaining structural simplicity and efficiency.& COPY; 2023 Published by Elsevier B.V.
引用
收藏
页数:13
相关论文
共 71 条
  • [1] An Overview of Evolutionary Algorithms for Parameter Optimization
    Baeck, Thomas
    Schwefel, Hans-Paul
    [J]. EVOLUTIONARY COMPUTATION, 1993, 1 (01) : 1 - 23
  • [2] Baker B., 2016, ARXIV
  • [3] Baker B, 2017, ARXIV
  • [4] Baldeon Calisto MariaG., 2020, Medical Imaging 2020: Image Processing., V11313, P459
  • [5] AdaResU-Net: Multiobjective adaptive convolutional neural network for medical image segmentation
    Baldeon-Calisto, Maria
    Lai-Yuen, Susana K.
    [J]. NEUROCOMPUTING, 2020, 392 : 325 - 340
  • [6] Bender G, 2018, PR MACH LEARN RES, V80
  • [7] Brock Andrew, 2018, PREPRINT
  • [8] Cai H, 2018, AAAI CONF ARTIF INTE, P2787
  • [9] Xception: Deep Learning with Depthwise Separable Convolutions
    Chollet, Francois
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 1800 - 1807
  • [10] ChamNet: Towards Efficient Network Design through Platform-Aware Model Adaptation
    Dai, Xiaoliang
    Zhang, Peizhao
    Wu, Bichen
    Yin, Hongxu
    Sun, Fei
    Wang, Yanghan
    Dukhan, Marat
    Hu, Yunqing
    Wu, Yiming
    Jia, Yangqing
    Vajda, Peter
    Uyttendaele, Matt
    Jha, Niraj K.
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 11390 - 11399