State-Space Closure: Revisiting Endless Online Level Generation via Reinforcement Learning

被引:0
|
作者
Wang, Ziqi [1 ,2 ]
Shu, Tianye [1 ,2 ]
Liu, Jialin [1 ,2 ]
机构
[1] Southern Univ Sci & Technol SUSTech, Dept Comp Sci & Engn, Guangdong Prov Key Lab Brain Inspired Intelligent, Shenzhen 518055, Peoples R China
[2] Southern Univ Sci & Technol SUSTech, Res Inst Trustworthy Autonomous Syst, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Games; Training; Reinforcement learning; Generators; Deep learning; Visualization; Hamming distances; Content diversity; online level generation (OLG); platformer games; procedural content generation (PCG); PCG via reinforcement learning (RL);
D O I
10.1109/TG.2023.3262297
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this letter, we revisit endless online level generation with the recently proposed experience-driven procedural content generation via reinforcement learning (EDRL) framework. Inspired by an observation that EDRL tends to generate recurrent patterns, we formulate a notion of state-space closure, which makes any stochastic state appeared possibly in an infinite-horizon online generation process, that can be found within a finite horizon. Through theoretical analysis, we find that even though state-space closure arises a concern about diversity, it generalizes EDRL trained with a finite horizon to the infinite-horizon scenario without deterioration of content quality. Moreover, we verify the quality and the diversity of contents generated by EDRL via empirical studies on the widely used Super Mario Bros. benchmark. Experimental results reveal that the diversity of levels generated by EDRL is limited due to the state-space closure, whereas their quality does not deteriorate in a horizon that is longer than the one specified in the training. Concluding our outcomes and analysis, future work on endless online level generation via reinforcement learning should address the issue of diversity while assuring the occurrence of state-space closure and quality.
引用
收藏
页码:489 / 492
页数:4
相关论文
共 50 条
  • [31] Online identification and control of PDEs via reinforcement learning methods
    Alla, Alessandro
    Pacifico, Agnese
    Palladino, Michele
    Pesare, Andrea
    ADVANCES IN COMPUTATIONAL MATHEMATICS, 2024, 50 (04)
  • [32] A Computational Framework for Automatic Online Path Generation of Robotic Inspection Tasks via Coverage Planning and Reinforcement Learning
    Jing, Wei
    Goh, Chun Fan
    Rajaraman, Mabaran
    Gao, Fei
    Park, Sooho
    Liu, Yong
    Shimada, Kenji
    IEEE ACCESS, 2018, 6 : 54854 - 54864
  • [33] Leveraging Reinforcement Learning and WaveFunctionCollapse for Improved Procedural Level Generation
    Babin, Mathias
    Katchabaw, Michael
    PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON THE FOUNDATIONS OF DIGITAL GAMES, FDG 2021, 2021,
  • [34] Detecting State of Charge False Reporting Attacks via Reinforcement Learning Approach
    Alomrani, Mhd Ali
    Tushar, Mosaddek Hossain Kamal
    Kundur, Deepa
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (10) : 10467 - 10476
  • [35] A deep learning method for pricing high-dimensional American-style options via state-space partition
    Han, Yuecai
    Zheng, Xudong
    COMPUTATIONAL & APPLIED MATHEMATICS, 2024, 43 (03)
  • [36] The Applicability of Reinforcement Learning for the Automatic Generation of State Preparation Circuits
    Gabor, Thomas
    Zorn, Maximilian
    Linnhoff-Popien, Claudia
    PROCEEDINGS OF THE 2022 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2022, 2022, : 2196 - 2204
  • [37] Chinese Emotional Dialogue Response Generation via Reinforcement Learning
    Lan, Rushi
    Wang, Jing
    Huang, Wenming
    Deng, Zhenrong
    Sun, Xiyan
    Chen, Zhuo
    Luo, Xiaonan
    ACM TRANSACTIONS ON INTERNET TECHNOLOGY, 2021, 21 (04)
  • [38] State-space model with deep learning for functional dynamics estimation in resting-state fMRI
    Suk, Heung-Il
    Wee, Chong-Yaw
    Lee, Seong-Whan
    Shen, Dinggang
    NEUROIMAGE, 2016, 129 : 292 - 307
  • [39] Accelerating Deep Reinforcement Learning via Phase-Level Parallelism for Robotics Applications
    Kim, Yang-Gon
    Han, Yun-Ki
    Shin, Jae-Kang
    Kim, Jun-Kyum
    Kim, Lee-Sup
    IEEE COMPUTER ARCHITECTURE LETTERS, 2024, 23 (01) : 41 - 44
  • [40] Data-driven control of wind turbine under online power strategy via deep learning and reinforcement learning
    Li, Tenghui
    Yang, Jin
    Ioannou, Anastasia
    RENEWABLE ENERGY, 2024, 234