Exploiting semantic segmentation to boost reinforcement learning in video game environments

被引:1
|
作者
Montalvo, Javier [1 ]
Garcia-Martin, Alvaro [1 ]
Bescos, Jesus [1 ]
机构
[1] Univ Autonoma Madrid, VPULab, Ciudad Univ Cantoblanco, E-28049 Madrid, Spain
关键词
Semantic segmentation; Reinforcement learning; Domain adaptation; Synthetic data;
D O I
10.1007/s11042-022-13695-1
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this work we explore enhancing performance of reinforcement learning algorithms in video game environments by feeding it better, more relevant data. For this purpose, we use semantic segmentation to transform the images that would be used as input for the reinforcement learning algorithm from their original domain to a simplified semantic domain with just silhouettes and class labels instead of textures and colors, and then we train the reinforcement learning algorithm with these simplified images. We have conducted different experiments to study multiple aspects: feasibility of our proposal, and potential benefits to model generalization and transfer learning. Experiments have been performed with the Super Mario Bros video game as the testing environment. Our results show multiple advantages for this method. First, it proves that using semantic segmentation enables reaching higher performance than the baseline reinforcement learning algorithm without modifying the actual algorithm, and in fewer episodes; second, it shows noticeable performance improvements when training on multiple levels at the same time; and finally, it allows to apply transfer learning for models trained on visually different environments. We conclude that using semantic segmentation can certainly help reinforcement learning algorithms that work with visual data, by refining it. Our results also suggest that other computer vision techniques may also be beneficial for data prepossessing. Models and code will be available on github upon acceptance.
引用
收藏
页码:10961 / 10979
页数:19
相关论文
共 50 条
  • [31] A Review of Reinforcement Learning for Semantic Communications
    Xiao Yan
    Fan Xiumei
    Kok-Lim Alvin Yau
    Xie Zhixin
    Men Rui
    Yuan Gang
    Journal of Network and Systems Management, 2025, 33 (3)
  • [32] Reinforcement Learning in Card Game Environments Using Monte Carlo Methods and Artificial Neural Networks
    Baykal, Omer
    Alpaslan, Ferda Nur
    2019 4TH INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND ENGINEERING (UBMK), 2019, : 618 - 623
  • [33] How deep learning is empowering semantic segmentation Traditional and deep learning techniques for semantic segmentation: A comparison
    Sehar, Uroosa
    Naseem, Muhammad Luqman
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (21) : 30519 - 30544
  • [34] Application of reinforcement learning to the game of Othello
    van Eck, Nees Jan
    van Wezel, Michiel
    COMPUTERS & OPERATIONS RESEARCH, 2008, 35 (06) : 1999 - 2017
  • [35] Reinforcement learning rules in a repeated game
    Bell A.M.
    Computational Economics, 2001, 18 (01) : 89 - 110
  • [36] Social Reinforcement Learning in Game Playing
    Kiourt, Chairi
    Kalles, Dimitris
    2012 IEEE 24TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2012), VOL 1, 2012, : 322 - 326
  • [37] TwinsAdvNet : Adversarial Learning for Semantic Segmentation
    Wang, Dongli
    Wang, Bo
    Zhou, Yan
    2019 7TH IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (IEEE GLOBALSIP), 2019,
  • [38] Hierarchical Contrastive Learning for Semantic Segmentation
    Jiang, Jie
    He, Xingjian
    Wang, Weining
    Lu, Hanqing
    Liu, Jing
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [39] Stereo SLAM in Dynamic Environments Using Semantic Segmentation
    Ai, Yongbao
    Sun, Qianchong
    Xi, Zhipeng
    Li, Na
    Dong, Jianmeng
    Wang, Xiang
    ELECTRONICS, 2023, 12 (14)
  • [40] Dual Correlation Network for Efficient Video Semantic Segmentation
    An, Shumin
    Liao, Qingmin
    Lu, Zongqing
    Xue, Jing-Hao
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (03) : 1572 - 1585