Rogue-Gym: A New Challenge for Generalization in Reinforcement Learning

被引:1
|
作者
Kanagawa, Yuji [1 ]
Kaneko, Tomoyuki [2 ]
机构
[1] Univ Tokyo, Grad Sch Arts & Sci, Tokyo, Japan
[2] Univ Tokyo, Interfac Initiat Informat Studies, Tokyo, Japan
来源
2019 IEEE CONFERENCE ON GAMES (COG) | 2019年
关键词
roguelike games; reinforcement learning; generalization; domain adaptation; neural networks; ENVIRONMENT;
D O I
10.1109/cig.2019.8848075
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose Rogue-Gym, a simple and classic style roguelike game built for evaluating generalization in reinforcement learning (RL). Combined with the recent progress of deep neural networks, RL has successfully trained human-level agents without human knowledge in many games such as those for Atari 2600. However, it has been pointed out that agents trained with RL methods often overfit the training environment, and they work poorly in slightly different environments. To investigate this problem, some research environments with procedural content generation have been proposed. Following these studies, we propose the use of roguelikes as a benchmark for evaluating the generalization ability of RL agents. In our Rogue-Gym, agents need to explore dungeons that are structured differently each time they start a new game. Thanks to the very diverse structures of the dungeons, we believe that the generalization benchmark of Rogue-Gym is sufficiently fair. In our experiments, we evaluate a standard reinforcement learning method, PPO, with and without enhancements for generalization. The results show that some enhancements believed to be effective fail to mitigate the overfitting in Rogue-Gym, although others slightly improve the generalization ability.
引用
收藏
页数:8
相关论文
共 50 条
  • [21] Scaling, Control and Generalization in Reinforcement Learning Level Generators
    Earle, Sam
    Jiang, Zehua
    Togelius, Julian
    2024 IEEE CONFERENCE ON GAMES, COG 2024, 2024,
  • [22] Federated reinforcement learning for robot motion planning with zero-shot generalization
    Yuan, Zhenyuan
    Xu, Siyuan
    Zhu, Minghui
    AUTOMATICA, 2024, 166
  • [23] BaziGooshi: A Hybrid Model of Reinforcement Learning for Generalization in Gameplay
    Karimi, Sara
    Asadi, Sahar
    Payberah, Amir H.
    IEEE TRANSACTIONS ON GAMES, 2024, 16 (03) : 722 - 734
  • [24] Reinforcement Learning in Multi-agent Games: Open AI Gym Diplomacy Environment
    Cruz, Diogo
    Cruz, Jose Aleixo
    Cardoso, Henrique Lopes
    PROGRESS IN ARTIFICIAL INTELLIGENCE, EPIA 2019, PT I, 2019, 11804 : 49 - 60
  • [25] Improving reinforcement learning with human assistance: an argument for human subject studies with HIPPO Gym
    Matthew E. Taylor
    Nicholas Nissen
    Yuan Wang
    Neda Navidi
    Neural Computing and Applications, 2023, 35 : 23429 - 23439
  • [26] Improving reinforcement learning with human assistance: an argument for human subject studies with HIPPO Gym
    Taylor, Matthew E.
    Nissen, Nicholas
    Wang, Yuan
    Navidi, Neda
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (32) : 23429 - 23439
  • [27] Time Horizon Generalization in Reinforcement Learning: Generalizing Multiple Q-Tables in Q-Learning Agents
    Hatcho, Yasuyo
    Hattori, Kiyohiko
    Takadama, Keiki
    JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2009, 13 (06) : 667 - 674
  • [28] Solution of an optimal routing problem by reinforcement learning with generalization ability
    Iima H.
    Oonishi H.
    IEEJ Transactions on Electronics, Information and Systems, 2019, 139 (12) : 1494 - 1500
  • [29] Metrics for Assessing Generalization of Deep Reinforcement Learning in Parameterized Environments
    Aleksandrowicz, Maciej
    Jaworek-Korjakowska, Joanna
    JOURNAL OF ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING RESEARCH, 2024, 14 (01) : 45 - 61
  • [30] IMPROVING GENERALIZATION OF REINFORCEMENT LEARNING USING A BILINEAR POLICY NETWORK
    Fang, Fen
    Liang, Wenyu
    Wu, Yan
    Xu, Qianli
    Lim, Joo-Hwee
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 991 - 995