An Algorithmic Framework for the Optimization of Deep Neural Networks Architectures and Hyperparameters

被引:0
作者
Keisler, Julie [1 ,2 ,3 ]
Talbi, El-Ghazali [2 ,3 ]
Claudel, Sandra [1 ,4 ]
Cabriel, Gilles [1 ,4 ]
机构
[1] EDF Lab Paris Saclay, Bd Gaspard Monge, F-91120 Palaiseau, France
[2] Univ Lille, 170 Av Bretagne, F-59000 Lille, France
[3] INRIA, 170 Av Bretagne, F-59000 Lille, France
[4] Univ Lille, EGID, U1011, F-59000 Lille, France
关键词
neural architecture search; hyperparameters optimization; metaheuristics; evolutionary algorithm; time series forecasting;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we propose DRAGON (for DiRected Acyclic Graph OptimizatioN), an algorithmic framework to automatically generate efficient deep neural networks architectures and optimize their associated hyperparameters. The framework is based on evolving Directed Acyclic Graphs (DAGs), defining a more flexible search space than the existing ones in the literature. It allows mixtures of different classical operations: convolutions, recurrences and dense layers, but also more newfangled operations such as self-attention. Based on this search space we propose neighbourhood and evolution search operators to optimize both the architecture and hyper-parameters of our networks. These search operators can be used with any metaheuristic capable of handling mixed search spaces. We tested our algorithmic framework with an asynchronous evolutionary algorithm on a time series forecasting benchmark. The results demonstrate that DRAGON outperforms state-of-theart handcrafted models and AutoML techniques for time series forecasting on numerous datasets. DRAGON has been implemented as a python open-source package1.
引用
收藏
页数:33
相关论文
共 56 条
  • [1] Abu-Aisheh Zeina, 2015, 4th International Conference on Pattern Recognition Applications and Methods (ICPRAM 2015). Proceedings, P271
  • [2] Evolutionary synthesis of logic circuits using Information Theory
    Aguirre, AH
    Coello, CAC
    [J]. ARTIFICIAL INTELLIGENCE REVIEW, 2003, 20 (3-4) : 445 - 471
  • [3] Review of ML and AutoML Solutions to Forecast Time-Series Data
    Alsahref, Ahmad
    Aggarwal, Karan
    Sonia
    Kumar, Manoj
    Mishra, Ashutosh
    [J]. ARCHIVES OF COMPUTATIONAL METHODS IN ENGINEERING, 2022, 29 (07) : 5297 - 5311
  • [4] DENSER: deep evolutionary network structured representation
    Assuncao, Filipe
    Lourenco, Nuno
    Machado, Penousal
    Ribeiro, Bernardete
    [J]. GENETIC PROGRAMMING AND EVOLVABLE MACHINES, 2019, 20 (01) : 5 - 35
  • [5] Bender G, 2018, PR MACH LEARN RES, V80
  • [6] Brock A, 2017, Arxiv, DOI arXiv:1708.05344
  • [7] Bayesian neural architecture search using a training-free performance metric
    Camero, Andres
    Wang, Hao
    Alba, Enrique
    Back, Thomas
    [J]. APPLIED SOFT COMPUTING, 2021, 106
  • [8] Graph-based Neural Architecture Search with Operation Embeddings
    Chatzianastasis, Michail
    Dasoulas, George
    Siolas, Georgios
    Vazirgiannis, Michalis
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, : 393 - 402
  • [9] Chen D., 2021, arXiv
  • [10] Cho KYHY, 2014, Arxiv, DOI [arXiv:1406.1078, DOI 10.48550/ARXIV.1406.1078]