Surrogate-assisted evolutionary neural architecture search with network embedding

被引:5
作者
Fan, Liang [1 ]
Wang, Handing [1 ]
机构
[1] Xidian Univ, Sch Artificial Intelligence, Xian 710071, Shaanxi, Peoples R China
基金
中国国家自然科学基金;
关键词
Neural architecture search; Evolutionary algorithm; Surrogate-assisted; Network embedding; PARTICLE SWARM OPTIMIZATION; ALGORITHM;
D O I
10.1007/s40747-022-00929-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To accelerate the performance estimation in neural architecture search, recently proposed algorithms adopt surrogate models to predict the performance of neural architectures instead of training the network from scratch. However, it is time-consuming to collect sufficient labeled architectures for surrogate model training. To enhance the capability of surrogate models using a small amount of training data, we propose a surrogate-assisted evolutionary algorithm with network embedding for neural architecture search (SAENAS-NE). Here, an unsupervised learning method is used to generate meaningful representation of each architecture and the architectures with more similar structures are closer in the embedding space, which considerably benefits the training of surrogate models. In addition, a new environmental selection based on a reference population is designed to keep diversity of the population in each generation and an infill criterion for handling the trade-off between convergence and model uncertainty is proposed for re-evaluation. Experimental results on three different NASBench and DARTS search space illustrate that network embedding makes the surrogate model achieve comparable or superior performance. The superiority of our proposed method SAENAS-NE over other state-of-the-art neural architecture algorithm has been verified in the experiments.
引用
收藏
页码:3313 / 3331
页数:19
相关论文
共 79 条
[1]  
Awad N., 2020, ARXIV
[2]  
Baker B, 2017, ARXIV
[3]  
Bender G, 2018, PR MACH LEARN RES, V80
[4]   Large-Scale Machine Learning with Stochastic Gradient Descent [J].
Bottou, Leon .
COMPSTAT'2010: 19TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL STATISTICS, 2010, :177-186
[5]  
Brock A, 2017, ARXIV
[6]  
Burges C, 2005, P 22 INT C MACH LEAR, P89, DOI [DOI 10.1145/1102351.1102363, 10.1145/1102351.1102363]
[7]  
Byla E, 2019, ARXIV
[8]  
Cheng Hao, 2020, ARXIV
[9]  
Chrabaszcz P., 2017, ARXIV
[10]  
Chu X., 2019, Fairnas: Rethinking evaluation fairness of weight sharing neural architecture search