TY - JOUR
T1 - Experienced Gray Wolf Optimization Through Reinforcement Learning and Neural Networks
AU - Emary, E.
AU - Zawbaa, Hossam M.
AU - Grosan, Crina
N1 - Funding Information:
Manuscript received May 26, 2016; revised September 25, 2016; accepted November 25, 2016. Date of publication January 10, 2017; date of current version February 15, 2018. This work was supported in part by the IPROCOM Marie Curie initial training network, funded through the People Programme (Marie Curie Actions) of the European Union’s Seventh Framework Programme FP7/2007-2013/ under REA Grant 316555, and in part by the Romanian National Authority for Scientific Research, CNDI-UEFISCDI, under Project PN-II-PT-PCCA-2011-3.2-0917. (Corresponding author: Hossam M. Zawbaa.) E. Emary is with the Faculty of Computers and Information, Cairo University, Giza 12613, Egypt.
Funding Information:
This work was supported in part by the IPROCOM Marie Curie initial training network, funded through the People Programme (Marie Curie Actions) of the European Union's Seventh Framework Programme FP7/2007-2013/under REA Grant 316555, and in part by the Romanian National Authority for Scientific Research, CNDI-UEFISCDI, under Project PN-II-PT-PCCA-2011-3.2-0917.
Publisher Copyright:
© 2016 IEEE.
PY - 2018/3/1
Y1 - 2018/3/1
N2 - In this paper, a variant of gray wolf optimization (GWO) that uses reinforcement learning principles combined with neural networks to enhance the performance is proposed. The aim is to overcome, by reinforced learning, the common challenge of setting the right parameters for the algorithm. In GWO, a single parameter is used to control the exploration/exploitation rate, which influences the performance of the algorithm. Rather than using a global way to change this parameter for all the agents, we use reinforcement learning to set it on an individual basis. The adaptation of the exploration rate for each agent depends on the agent's own experience and the current terrain of the search space. In order to achieve this, experience repository is built based on the neural network to map a set of agents' states to a set of corresponding actions that specifically influence the exploration rate. The experience repository is updated by all the search agents to reflect experience and to enhance the future actions continuously. The resulted algorithm is called experienced GWO (EGWO) and its performance is assessed on solving feature selection problems and on finding optimal weights for neural networks algorithm. We use a set of performance indicators to evaluate the efficiency of the method. Results over various data sets demonstrate an advance of the EGWO over the original GWO and over other metaheuristics, such as genetic algorithms and particle swarm optimization.
AB - In this paper, a variant of gray wolf optimization (GWO) that uses reinforcement learning principles combined with neural networks to enhance the performance is proposed. The aim is to overcome, by reinforced learning, the common challenge of setting the right parameters for the algorithm. In GWO, a single parameter is used to control the exploration/exploitation rate, which influences the performance of the algorithm. Rather than using a global way to change this parameter for all the agents, we use reinforcement learning to set it on an individual basis. The adaptation of the exploration rate for each agent depends on the agent's own experience and the current terrain of the search space. In order to achieve this, experience repository is built based on the neural network to map a set of agents' states to a set of corresponding actions that specifically influence the exploration rate. The experience repository is updated by all the search agents to reflect experience and to enhance the future actions continuously. The resulted algorithm is called experienced GWO (EGWO) and its performance is assessed on solving feature selection problems and on finding optimal weights for neural networks algorithm. We use a set of performance indicators to evaluate the efficiency of the method. Results over various data sets demonstrate an advance of the EGWO over the original GWO and over other metaheuristics, such as genetic algorithms and particle swarm optimization.
KW - Adaptive exploration rate
KW - artificial neural network (ANN)
KW - experienced gray Wolf optimization (EGWO)
KW - gray Wolf optimization (GWO)
KW - reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85009932318&partnerID=8YFLogxK
U2 - 10.1109/TNNLS.2016.2634548
DO - 10.1109/TNNLS.2016.2634548
M3 - Article
C2 - 28092578
AN - SCOPUS:85009932318
SN - 2162-237X
VL - 29
SP - 681
EP - 694
JO - IEEE Transactions on Neural Networks and Learning Systems
JF - IEEE Transactions on Neural Networks and Learning Systems
IS - 3
ER -