Differential evolution with mixed mutation strategy based on deep reinforcement learning

被引:49
作者
Tan, Zhiping [1 ]
Li, Kangshun [2 ]
机构
[1] Guangdong Polytech Normal Univ, Coll Elect & Informat, Guangzhou 510665, Peoples R China
[2] South China Agr Univ, Coll Math & Informat, Guangzhou 510642, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Differential evolution; Mixed mutation strategy; Fitness landscape; Deep reinforcement learning; Deep Q-learning; ALGORITHM; OPTIMIZATION; ENSEMBLE; PARAMETERS; OPERATOR; DESIGN;
D O I
10.1016/j.asoc.2021.107678
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The performance of differential evolution (DE) algorithm significantly depends on mutation strategy. However, there are six commonly used mutation strategies in DE. It is difficult to select a reasonable mutation strategy in solving the different real-life optimization problems. In general, the selection of the most appropriate mutation strategy is based on personal experience. To address this problem, a mixed mutation strategy DE algorithm based on deep Q-network (DQN), named DEDQN is proposed in this paper, in which a deep reinforcement learning approach realizes the adaptive selection of mutation strategy in the evolution process. Two steps are needed for the application of DQN to DE. First, the DQN is trained offline through collecting the data about fitness landscape and the benefit (reward) of applying each mutation strategy during multiple runs of DEDQN tackling the training functions. Second, the mutation strategy is predicted by the trained DQN at each generation according to the fitness landscape of every test function. Besides, a historical memory parameter adaptation mechanism is also utilized to improve the DEDQN. The performance of the DEDQN algorithm is evaluated by the CEC2017 benchmark function set, and five state-of-the-art DE algorithms are compared with the DEDQN in the experiments. The experimental results indicate the competitive performance of the proposed algorithm. (C) 2021 Published by Elsevier B.V.
引用
收藏
页数:13
相关论文
共 50 条
[31]   MDE: Differential Evolution with Merit-based Mutation Strategy [J].
Ibrahim, Amin ;
Rahnamayan, Shahryar ;
Martin, Miguel Vargas .
2014 IEEE SYMPOSIUM ON DIFFERENTIAL EVOLUTION (SDE), 2014, :41-48
[32]   Enhanced Mutation Strategy for Differential Evolution [J].
Kumar, Pravesh ;
Pant, Millie .
2012 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2012,
[33]   Differential Evolution with Improved Mutation Strategy [J].
Wan, Shuzhen ;
Xiong, Shengwu ;
Kou, Jialiang ;
Liu, Yi .
ADVANCES IN SWARM INTELLIGENCE, PT I, 2011, 6728 :431-438
[34]   Network Resource Allocation Strategy Based on Deep Reinforcement Learning [J].
Zhang, Shidong ;
Wang, Chao ;
Zhang, Junsan ;
Duan, Youxiang ;
You, Xinhong ;
Zhang, Peiying .
IEEE OPEN JOURNAL OF THE COMPUTER SOCIETY, 2020, 1 (01) :86-94
[35]   A directional mutation operator for differential evolution algorithms [J].
Zhang, Xin ;
Yuen, Shiu Yin .
APPLIED SOFT COMPUTING, 2015, 30 :529-548
[36]   Improving Dual-Population Differential Evolution Based on Hierarchical Mutation and Selection Strategy [J].
Huang, Yawei ;
Qian, Xuezhong ;
Song, Wei .
ELECTRONICS, 2024, 13 (01)
[37]   Landscape-based adaptive operator selection mechanism for differential evolution [J].
Sallam, Karam M. ;
Elsayed, Saber M. ;
Sarker, Ruhul A. ;
Essam, Daryl L. .
INFORMATION SCIENCES, 2017, 418 :383-404
[38]   Enhanced differential evolution with hierarchical selection mutation and distance-based selection strategy [J].
Luo, Zhenyong ;
Qian, Xuezhong ;
Song, Wei .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 144
[39]   Reinforcement Learning for Dynamic Mutation Process Control in Multi-Objective Differential Evolution [J].
Tao, Lue ;
Wang, Gongshu ;
Yang, Yang ;
Dong, Yun ;
Su, Lijie .
IFAC PAPERSONLINE, 2022, 55 (15) :117-122
[40]   Trigonometric mutation and successful-parent-selection based adaptive asynchronous differential evolution [J].
Yadav, Vaishali ;
Yadav, Ashwani Kumar ;
Kaur, Manjit ;
Singh, Dilbag .
JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2021, 13 (12) :5829-5846