Deep Reinforcement Learning in Strategic Board Game Environments

被引:9
作者
Xenou, Konstantia [1 ]
Chalkiadakis, Georgios [1 ]
Afantenos, Stergos [2 ]
机构
[1] Tech Univ Crete, Sch Elect & Comp Engn, Khania, Greece
[2] Univ Paul Sabatier, Inst Rech Informat Toulouse IRIT, Toulouse, France
来源
MULTI-AGENT SYSTEMS, EUMAS 2018 | 2019年 / 11450卷
关键词
Deep Reinforcement Learning; Strategic board games;
D O I
10.1007/978-3-030-14174-5_16
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper we propose a novel Deep Reinforcement Learning (DRL) algorithm that uses the concept of "action-dependent state features", and exploits it to approximate the Q-values locally, employing a deep neural network with parallel Long Short Term Memory (LSTM) components, each one responsible for computing an action-related Q-value. As such, all computations occur simultaneously, and there is no need to employ "target" networks and experience replay, which are techniques regularly used in the DRL literature. Moreover, our algorithm does not require previous training experiences, but trains itself online during game play. We tested our approach in the Settlers Of Catan multi-player strategic board game. Our results confirm the effectiveness of our approach, since it outperforms several competitors, including the state-of-the-art jSettler heuristic algorithm devised for this particular domain.
引用
收藏
页码:233 / 248
页数:16
相关论文
共 28 条
  • [1] Afantenos Stergos D., 2015, P 2015 C EMP METH NA, P928
  • [2] Anschel O., 2016, arXiv
  • [3] Bellman R, 2013, DYNAMIC PROGRAMMING
  • [4] Cuayahuitl H., 2015, P NIPS DEEP REINFORC
  • [5] Dearden R, 1998, FIFTEENTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-98) AND TENTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICAL INTELLIGENCE (IAAI-98) - PROCEEDINGS, P761
  • [6] Dobre MS, 2015, IEEE CONF COMPU INTE, P60, DOI 10.1109/CIG.2015.7317942
  • [7] Finnman P, 2016, Deep reinforcement learning compared with Q-table learning applied to backgammon
  • [8] Graves A, 2012, STUD COMPUT INTELL, V385, P1, DOI [10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]
  • [9] Guhe Markus., 2014, 2014 IEEE Conference on Computational Intelligence and Games, P1
  • [10] Hausknecht M, 2015, 2015 AAAI FALL S SER, P29, DOI 10.1.1.696.1421