Experienced Gray Wolf Optimization Through Reinforcement Learning and Neural Networks

被引:154
作者
Emary, E. [1 ]
Zawbaa, Hossam M. [2 ,3 ]
Grosan, Crina [3 ,4 ]
机构
[1] Cairo Univ, Fac Comp & Informat, Giza 12613, Egypt
[2] Beni Suef Univ, Fac Comp & Informat, Bani Suwayf 62511, Egypt
[3] Babes Bolyai Univ, Fac Math & Comp Sci, Cluj Napoca 400084, Romania
[4] Brunel Univ, Coll Engn Design & Phys Sci, Uxbridge UB8 3PH, Middx, England
关键词
Adaptive exploration rate; artificial neural network (ANN); experienced gray wolf optimization (EGWO); gray wolf optimization (GWO); reinforcement learning; PARTICLE SWARM OPTIMIZATION; FEATURE-SELECTION; ALGORITHM;
D O I
10.1109/TNNLS.2016.2634548
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, a variant of gray wolf optimization (GWO) that uses reinforcement learning principles combined with neural networks to enhance the performance is proposed. The aim is to overcome, by reinforced learning, the common challenge of setting the right parameters for the algorithm. In GWO, a single parameter is used to control the exploration/exploitation rate, which influences the performance of the algorithm. Rather than using a global way to change this parameter for all the agents, we use reinforcement learning to set it on an individual basis. The adaptation of the exploration rate for each agent depends on the agent's own experience and the current terrain of the search space. In order to achieve this, experience repository is built based on the neural network to map a set of agents' states to a set of corresponding actions that specifically influence the exploration rate. The experience repository is updated by all the search agents to reflect experience and to enhance the future actions continuously. The resulted algorithm is called experienced GWO (EGWO) and its performance is assessed on solving feature selection problems and on finding optimal weights for neural networks algorithm. We use a set of performance indicators to evaluate the efficiency of the method. Results over various data sets demonstrate an advance of the EGWO over the original GWO and over other metaheuristics, such as genetic algorithms and particle swarm optimization.
引用
收藏
页码:681 / 694
页数:14
相关论文
共 32 条
[1]  
[Anonymous], PATTERN RECOGNITION
[2]  
[Anonymous], 2015, Reinforcement Learning: An Introduction
[3]  
[Anonymous], 1973, Pattern Classification and Scene Analysis
[4]  
[Anonymous], 2016, T AUTOM SCI ENG, DOI DOI 10.1038/ONCSIS.2016.35
[5]   Improved binary PSO for feature selection using gene expression data [J].
Chuang, Li-Yeh ;
Chang, Hsueh-Wei ;
Tu, Chung-Jui ;
Yang, Cheng-Hong .
COMPUTATIONAL BIOLOGY AND CHEMISTRY, 2008, 32 (01) :29-38
[6]   Improved binary particle swarm optimization using catfish effect for feature selection [J].
Chuang, Li-Yeh ;
Tsai, Sheng-Wei ;
Yang, Cheng-Hong .
EXPERT SYSTEMS WITH APPLICATIONS, 2011, 38 (10) :12699-12707
[7]   Deep Direct Reinforcement Learning for Financial Signal Representation and Trading [J].
Deng, Yue ;
Bao, Feng ;
Kong, Youyong ;
Ren, Zhiquan ;
Dai, Qionghai .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (03) :653-664
[8]  
Eiben AE, 1994, LECT NOTES COMPUT SC, V866, P78
[9]   A two-stage gene selection scheme utilizing MRMR filter and GA wrapper [J].
El Akadi, Ali ;
Amine, Aouatif ;
El Ouardighi, Abdeljalil ;
Aboutajdine, Driss .
KNOWLEDGE AND INFORMATION SYSTEMS, 2011, 26 (03) :487-500
[10]   Binary grey wolf optimization approaches for feature selection [J].
Emary, E. ;
Zawba, Hossam M. ;
Hassanien, Aboul Ella .
NEUROCOMPUTING, 2016, 172 :371-381