A data-driven approach to neural architecture search initialization

被引:0
作者
Traore, Kalifou Rene [1 ,2 ]
Camero, Andres [2 ,3 ]
Zhu, Xiao Xiang [1 ,2 ]
机构
[1] Tech Univ Munich, Data Sci Earth Observat, Arcisstr 21, D-80333 Munich, Bavaria, Germany
[2] German Aerosp Ctr DLR, Remote Sensing Inst, Munchener Str 20, D-82234 Wessling, Bavaria, Germany
[3] Helmholtz AI, Munich, Germany
基金
欧洲研究理事会;
关键词
AutoML; Neural architecture search; Evolutionary computation; Search; Initialization; 68Txx; NETWORKS; POPULATION;
D O I
10.1007/s10472-022-09823-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Algorithmic design in neural architecture search (NAS) has received a lot of attention, aiming to improve performance and reduce computational cost. Despite the great advances made, few authors have proposed to tailor initialization techniques for NAS. However, the literature shows that a good initial set of solutions facilitates finding the optima. Therefore, in this study, we propose a data-driven technique to initialize a population-based NAS algorithm. First, we perform a calibrated clustering analysis of the search space, and second, we extract the centroids and use them to initialize a NAS algorithm. We benchmark our proposed approach against random and Latin hypercube sampling initialization using three population-based algorithms, namely a genetic algorithm, an evolutionary algorithm, and aging evolution, on CIFAR-10. More specifically, we use NAS-Bench-101 to leverage the availability of NAS benchmarks. The results show that compared to random and Latin hypercube sampling, the proposed initialization technique enables achieving significant long-term improvements for two of the search baselines, and sometimes in various search scenarios (various training budget). Besides, we also investigate how an initial population gathered on the tabular benchmark can be used for improving search on another dataset, the So2Sat LCZ-42. Our results show similar improvements on the target dataset, despite a limited training budget. Moreover, we analyse the distributions of solutions obtained and find that that the population provided by the data-driven initialization technique enables retrieving local optima (maxima) of high fitness and similar configurations.
引用
收藏
页码:413 / 440
页数:28
相关论文
共 59 条
[1]  
Alba E., 1993, New Trends in Neural Computation. International Workshop on Artificial Neural Networks. IWANN '93 Proceedings, P399
[2]  
Alba E., 1993, Artificial Neural Nets and Genetic Algorithms. Proceedings of the International Conference, P683
[3]  
Back Thomas., 1996, EVOLUTIONARY ALGORIT
[4]  
Brock A., 2017, ARXIV
[5]  
Camero A., 2020, ENG APPL ARTIF INTEL, V103946, P96
[6]  
Camero A., 2018, ARXIV
[7]   Bayesian neural architecture search using a training-free performance metric [J].
Camero, Andres ;
Wang, Hao ;
Alba, Enrique ;
Back, Thomas .
APPLIED SOFT COMPUTING, 2021, 106
[8]   Comparing Deep Recurrent Networks Based on the MAE Random Sampling, a First Approach [J].
Camero, Andres ;
Toutouh, Jamal ;
Alba, Enrique .
ADVANCES IN ARTIFICIAL INTELLIGENCE, CAEPIA 2018, 2018, 11160 :24-33
[9]   Fast, Accurate and Lightweight Super-Resolution with Neural Architecture Search [J].
Chu, Xiangxiang ;
Zhang, Bo ;
Ma, Hailong ;
Xu, Ruijun ;
Li, Qingyuan .
2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, :59-64
[10]  
Clerc M., 2008, Initialisations for particle swarm optimisation