Self-adaptation of Neuroevolution Algorithms Using Reinforcement Learning

被引:0
作者
Kogan, Michael [1 ]
Karns, Joshua [1 ]
Desell, Travis [1 ]
机构
[1] Rochester Inst Technol, Rochester, NY 14623 USA
来源
APPLICATIONS OF EVOLUTIONARY COMPUTATION (EVOAPPLICATIONS 2022) | 2022年
关键词
Neuroevolution; Reinforcement learning; FALA;
D O I
10.1007/978-3-031-02462-7_29
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Selecting an appropriate neural architecture for a given dataset is an open problem in machine learning. Neuroevolution algorithms, such as NEAT, have shown great promise in automating this process. An extension of NEAT called EXAMM has demonstrated the capability to generate recurrent neural architectures for various time series datasets. At each iteration the evolutionary process is furthered using randomly selected mutation or crossover operations, which are chosen in accordance with pre-assigned probabilities. In this paper we present a self-adapting version of EXAMM that incorporates finite action-set learning automata (FALA), a reinforcement learning technique. FALA is used to dynamically adjust the aforementioned probabilities, thereby guiding the evolutionary process, while also significantly reducing the number of required hyperparameters. It is also demonstrated that this approach improves the performance of the generated networks with statistical significance. Furthermore, the evolution of the adapted probabilities is analyzed to gain further insight into the inner workings of EXAMM.
引用
收藏
页码:452 / 467
页数:16
相关论文
共 20 条
[1]   Parallelism and evolutionary algorithms [J].
Alba, E ;
Tomassini, M .
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2002, 6 (05) :443-462
[2]  
[Anonymous], 2003, Networks of Learning Automata: Techniques for Online Stochastic Optimization
[3]  
[Anonymous], 2017, 2017 ANN IEEE INT SY
[4]   Autonomous Construction of Multiple Structures Using Learning Automata: Description and Experimental Validation [J].
Barros dos Santos, Sergio R. ;
Givigi, Sidney N., Jr. ;
Nascimento, Cairo L., Jr. .
IEEE SYSTEMS JOURNAL, 2015, 9 (04) :1376-1387
[5]  
Desell T., 2020, EMPIRICAL EXPLORATIO, P546
[6]   Large Scale Evolution of Convolutional Neural Networks Using Volunteer Computing [J].
Desell, Travis .
PROCEEDINGS OF THE 2017 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION (GECCO'17 COMPANION), 2017, :127-128
[7]   GPT-3: Its Nature, Scope, Limits, and Consequences [J].
Floridi, Luciano ;
Chiriatti, Massimo .
MINDS AND MACHINES, 2020, 30 (04) :681-694
[8]   On-line PID tuning for engine idle-speed control using continuous action reinforcement learning automata [J].
Howell, MN ;
Best, MC .
CONTROL ENGINEERING PRACTICE, 2000, 8 (02) :147-154
[9]  
Jardine P., 2018, REINFORCEMENT LEARNI
[10]   Adaptive predictive control of a differential drive robot tuned with reinforcement learning [J].
Jardine, P. Travis ;
Kogan, Michael ;
Givigi, Sidney N. ;
Yousefi, Shahram .
INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2019, 33 (02) :410-423