Multi-Agent Coordination Profiles Through State Space Perturbations

被引:6
作者
Asher, Derrik E. [1 ]
Garber-Barron, Michael [2 ]
Rodriguez, Sebastian S. [3 ]
Zaroukian, Erin [1 ]
Waytowich, Nicholas R. [4 ]
机构
[1] CCDC Army Res Lab, Informat Sci, Adelphi, MD 20783 USA
[2] Cornell Univ, Comp Sci, Ithaca, NY USA
[3] Univ Illinois, Comp Sci, Urbana, IL 61801 USA
[4] CCDC Army Res Lab, Human Sci, Adelphi, MD USA
来源
2019 6TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND COMPUTATIONAL INTELLIGENCE (CSCI 2019) | 2019年
关键词
predator-prey pursuit; coordination; simulation experiments; perturbation analysis; coordination profiles;
D O I
10.1109/CSCI49370.2019.00051
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The current work utilized a multi-agent reinforcement learning (MARL) algorithm embedded in a continuous predator-prey pursuit simulation environment to measure and evaluate coordination between cooperating agents. In this simulation environment, it is generally assumed that successful performance for cooperative agents necessarily results in the emergence of coordination, but a clear quantitative demonstration of coordination in this environment still does not exist. The current work focuses on 1) detecting emergent coordination between cooperating agents in a multi-agent predator-prey simulation environment, and 2) showing coordination profiles between cooperating agents extracted from systematic state perturbations. This work introduces a method for detecting and comparing the typically 'black-box' behavioral solutions that result from emergent coordination in multi-agent learning spatial tasks with a shared goal. Comparing coordination profiles can provide insights into overlapping patterns that define how agents learn to interact in cooperative multi-agent environments. Similarly, this approach provides an avenue for measuring and framing agents to coordinate with humans. In this way, the present work looks towards understanding and creating artificial team-mates that will strive to coordinate optimally.
引用
收藏
页码:249 / 252
页数:4
相关论文
共 16 条
[1]  
[Anonymous], 2019, IEEE T EVOLUTIONARY
[2]  
Barton S. L., 2018, ARXIV180904918
[3]   Measuring Collaborative Emergent Behavior in Multi-agent Reinforcement Learning [J].
Barton, Sean L. ;
Waytowich, Nicholas R. ;
Zaroukian, Erin ;
Asher, Derrik E. .
HUMAN SYSTEMS ENGINEERING AND DESIGN, IHSED2018, 2019, 876 :422-427
[4]   Evaluating the Coordination of Agents in Multi-agent Reinforcement Learning [J].
Barton, Sean L. ;
Zaroukian, Erin ;
Asher, Derrik E. ;
Waytowich, Nicholas R. .
INTELLIGENT HUMAN SYSTEMS INTEGRATION 2019, 2019, 903 :765-770
[5]   Reinforcement learning framework for collaborative agents interacting with Soldiers in dynamic military contexts [J].
Barton, Sean L. ;
Asher, Derrik .
NEXT-GENERATION ANALYST VI, 2018, 10653
[6]  
Cheung V, 2016, OPENAI GYM
[7]  
Engesser T., 2017, ARXIV170302196
[8]  
Foerster J., 2016, P ADV NEUR INF PROC, P2137
[9]   Ten challenges for making automation a "team player" in joint human-agent activity [J].
Klein, G ;
Woods, DD ;
Bradshaw, JM ;
Hoffman, RR ;
Feltovich, PJ .
IEEE INTELLIGENT SYSTEMS, 2004, 19 (06) :91-95
[10]  
Lauer M., 2000, Proceedings of the Seventeenth International Conference on Machine Learning