State-Space Closure: Revisiting Endless Online Level Generation via Reinforcement Learning

被引:0
|
作者
Wang, Ziqi [1 ,2 ]
Shu, Tianye [1 ,2 ]
Liu, Jialin [1 ,2 ]
机构
[1] Southern Univ Sci & Technol SUSTech, Dept Comp Sci & Engn, Guangdong Prov Key Lab Brain Inspired Intelligent, Shenzhen 518055, Peoples R China
[2] Southern Univ Sci & Technol SUSTech, Res Inst Trustworthy Autonomous Syst, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Games; Training; Reinforcement learning; Generators; Deep learning; Visualization; Hamming distances; Content diversity; online level generation (OLG); platformer games; procedural content generation (PCG); PCG via reinforcement learning (RL);
D O I
10.1109/TG.2023.3262297
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this letter, we revisit endless online level generation with the recently proposed experience-driven procedural content generation via reinforcement learning (EDRL) framework. Inspired by an observation that EDRL tends to generate recurrent patterns, we formulate a notion of state-space closure, which makes any stochastic state appeared possibly in an infinite-horizon online generation process, that can be found within a finite horizon. Through theoretical analysis, we find that even though state-space closure arises a concern about diversity, it generalizes EDRL trained with a finite horizon to the infinite-horizon scenario without deterioration of content quality. Moreover, we verify the quality and the diversity of contents generated by EDRL via empirical studies on the widely used Super Mario Bros. benchmark. Experimental results reveal that the diversity of levels generated by EDRL is limited due to the state-space closure, whereas their quality does not deteriorate in a horizon that is longer than the one specified in the training. Concluding our outcomes and analysis, future work on endless online level generation via reinforcement learning should address the issue of diversity while assuring the occurrence of state-space closure and quality.
引用
收藏
页码:489 / 492
页数:4
相关论文
共 50 条
  • [41] Topological Visualization Method for Understanding the Landscape of Value Functions and Structure of the State Space in Reinforcement Learning
    Nakamura, Yuki
    Shibuya, Takeshi
    ICAART: PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, VOL 2, 2020, : 370 - 377
  • [42] Efficient Communication via Self-Supervised Information Aggregation for Online and Offline Multiagent Reinforcement Learning
    Guan, Cong
    Chen, Feng
    Yuan, Lei
    Zhang, Zongzhang
    Yu, Yang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [43] Multiagent reinforcement learning with the partly high-dimensional state space
    Department of Electrical and Computer Engineering, Nagoya Institute of Technology, Nagoya, 466-8555, Japan
    Syst Comput Jpn, 2006, 9 (22-31): : 22 - 31
  • [44] Remote sensing image caption generation via transformer and reinforcement learning
    Xiangqing Shen
    Bing Liu
    Yong Zhou
    Jiaqi Zhao
    Multimedia Tools and Applications, 2020, 79 : 26661 - 26682
  • [45] Remote sensing image caption generation via transformer and reinforcement learning
    Shen, Xiangqing
    Liu, Bing
    Zhou, Yong
    Zhao, Jiaqi
    MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (35-36) : 26661 - 26682
  • [46] Safe-State Enhancement Method for Autonomous Driving via Direct Hierarchical Reinforcement Learning
    Gu, Ziqing
    Gao, Lingping
    Ma, Haitong
    Li, Shengbo Eben
    Zheng, Sifa
    Jing, Wei
    Chen, Junbo
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (09) : 9966 - 9983
  • [47] Control of Quadrotor Drone with Partial State Observation via Reinforcement Learning
    Shan, Guangcun
    Zhang, Yinan
    Gao, Yong
    Wang, Tian
    Chen, Jianping
    2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 1965 - 1968
  • [48] Reinforcement Learning Method for Continuous State Space Based on Dynamic Neural Network
    Sun, Wei
    Wang, Xuesong
    Cheng, Yuhu
    2008 7TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-23, 2008, : 750 - 754
  • [49] Metric State Space Reinforcement Learning for a Vision-Capable Mobile Robot
    Zhumatiy, Viktor
    Gomez, Faustino
    Hutter, Marcus
    Schmidhuber, Juergen
    INTELLIGENT AUTONOMOUS SYSTEMS 9, 2006, : 272 - +
  • [50] Swarm Reinforcement Learning Methods for Problems with Continuous State-Action Space
    Iima, Hitoshi
    Kuroe, Yasuaki
    Emoto, Kazuo
    2011 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2011, : 2173 - 2180