Efficient Elitist Cooperative Evolutionary Algorithm for Multi-Objective Reinforcement Learning

被引:5
|
作者
Zhou, Dan [1 ]
Du, Jiqing [1 ]
Arai, Sachiyo [1 ]
机构
[1] Chiba Univ, Grad Sch Sci & Engn, Dept Urban Environm Syst, Div Earth & Environm Sci, Chiba 2638522, Japan
关键词
Pareto optimization; Statistics; Social factors; Underwater vehicles; Measurement; Q-learning; Uncertainty; Reinforcement learning; Multi-objective reinforcement learning; efficient; cooperative; Pareto front; elite archive; GENETIC ALGORITHM;
D O I
10.1109/ACCESS.2023.3272115
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Sequential decision-making problems with multiple objectives are known as multi-objective reinforcement learning. In these scenarios, decision-makers require a complete Pareto front that consists of Pareto optimal solutions. Such a front enables decision-makers to understand the relationship between objectives and make informed decisions from a broad range of solutions. However, existing methods may be unable to search for solutions in concave regions of the Pareto front or lack global optimization ability, leading to incomplete Pareto fronts. To address this issue, we propose an efficient elitist cooperative evolutionary algorithm that maintains both an evolving population and an elite archive. The elite archive uses cooperative operations with various genetic operators to guide the evolving population, resulting in efficient searches for Pareto optimal solutions. The experimental results on submarine treasure hunting benchmarks demonstrate the effectiveness of the proposed method in solving various multi-objective reinforcement learning problems and providing decision-makers with a set of trade-off solutions between travel time and treasure amount, enabling them to make flexible and informed decisions based on their preferences. Therefore, the proposed method has the potential to be a useful tool for implementing real-world applications.
引用
收藏
页码:43128 / 43139
页数:12
相关论文
共 50 条
  • [21] Track Learning Agent Using Multi-objective Reinforcement Learning
    Shah, Rushabh
    Ruparel, Vidhi
    Prabhu, Mukul
    D'mello, Lynette
    FOURTH CONGRESS ON INTELLIGENT SYSTEMS, VOL 1, CIS 2023, 2024, 868 : 27 - 40
  • [22] Multi-Objective Evolutionary Federated Learning
    Zhu, Hangyu
    Jin, Yaochu
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (04) : 1310 - 1322
  • [23] An efficient data evacuation strategy using multi-objective reinforcement learning
    Li, Xiaole
    APPLIED INTELLIGENCE, 2022, 52 (07) : 7498 - 7512
  • [24] A Multi-Objective Virtual Network Migration Algorithm Based on Reinforcement Learning
    Wang, Desheng
    Zhang, Weizhe
    Han, Xiao
    Lin, Junren
    Tian, Yu-Chu
    IEEE TRANSACTIONS ON CLOUD COMPUTING, 2023, 11 (02) : 2039 - 2056
  • [25] An efficient data evacuation strategy using multi-objective reinforcement learning
    Xiaole Li
    Applied Intelligence, 2022, 52 : 7498 - 7512
  • [26] Multi-objective evolutionary algorithm with prediction in the objective space
    Guerrero-Pena, Elaine
    Ribeiro Araujo, Aluizio Fausto
    INFORMATION SCIENCES, 2019, 501 : 293 - 316
  • [27] Multi-Objective Reinforcement Learning Algorithm and Its Improved Convergency Method
    Zhao Jin
    Zhang Huajun
    2011 6TH IEEE CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS (ICIEA), 2011, : 2438 - 2445
  • [28] Multi-Objective Service Composition Using Reinforcement Learning
    Moustafa, Ahmed
    Zhang, Minjie
    SERVICE-ORIENTED COMPUTING, ICSOC 2013, 2013, 8274 : 298 - 312
  • [29] Multi-objective vehicle following decision algorithm based on reinforcement learning
    Dend X.-H.
    Hou J.
    Tan G.-H.
    Wan B.-Y.
    Cao T.-T.
    Kongzhi yu Juece/Control and Decision, 2021, 36 (10): : 2497 - 2503
  • [30] A reinforcement learning-assisted multi-objective evolutionary algorithm for generating green change plans of complex products
    Zheng, Ruizhao
    Zhang, Yong
    Sun, Xiaoyan
    Yang, Lei
    Song, Xianfang
    APPLIED SOFT COMPUTING, 2025, 170