Coordinated Reinforcement Learning Agents in a Multi-Agent Virtual Environment

被引:6
作者
Sause, William [1 ]
机构
[1] Nova SE Univ, Grad Sch Comp & Informat Sci, Ft Lauderdale, FL 33314 USA
来源
2013 12TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2013), VOL 1 | 2013年
关键词
Reinforcement learning; virtual environments; intelligent agents;
D O I
10.1109/ICMLA.2013.46
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This research presents a framework for coordinating multiple intelligent agents within a single virtual environment. Coordination is accomplished via a "next available agent" scheme while learning is achieved through the use of the Q-learning and Sarsa temporal difference reinforcement learning algorithms. To assess the effectiveness of each learning algorithm, experiments were conducted that measured an agent's ability to learn tasks in a static and dynamic environment while using both a fixed (FEP) and variable (VEP) epsilon-greedy probability rate. Results show that Sarsa, on average, outperformed Q-learning in almost all experiments. Overall, VEP resulted in higher percentages of successes and optimal successes than FEP, and showed convergence to the optimal policy when measuring the average number of time steps per episode.
引用
收藏
页码:227 / 230
页数:4
相关论文
共 6 条
  • [1] Ertel W., 2011, Introduction to artificial intelligence
  • [2] MENEGATTI E, 2004, P INT C INT ROB SYST, P2743
  • [3] Merrick K., 2007, THESIS U SYDNEY AUST
  • [4] MERRICK KE, 2009, MOTIVATED REINFORCEM
  • [5] Sutton R.S., 2017, Introduction to reinforcement learning
  • [6] Zamstein L. M., 2006, FLOR C REC ADV ROB M