When Does Communication Learning Need Hierarchical Multi-Agent Deep Reinforcement Learning

被引:2
作者
Ossenkopf, Marie [1 ]
Jorgensen, Mackenzie [2 ]
Geihs, Kurt [1 ]
机构
[1] Univ Kassel, Distributed Syst Grp, Wilhelmshoeher Allee 73, D-34121 Kassel, Germany
[2] Villanova Univ, Comp Sci, Villanova, PA 19085 USA
关键词
Agent communication; deep reinforcement learning; hierarchical learning; multi-agent systems;
D O I
10.1080/01969722.2019.1677335
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-agent systems need to communicate to coordinate a shared task. We show that a recurrent neural network (RNN) can learn a communication protocol for coordination, even if the actions to coordinate are performed steps after the communication phase. We show that a separation of tasks with different temporal scale is necessary for successful learning. We contribute a hierarchical deep reinforcement learning model for multi-agent systems that separates the communication and coordination task from the action picking through a hierarchical policy. We further on show, that a separation of concerns in communication is beneficial but not necessary. As a testbed, we propose the Dungeon Lever Game and we extend the Differentiable Inter-Agent Learning (DIAL) framework. We present and compare results from different model variations on the Dungeon Lever Game.
引用
收藏
页码:672 / 692
页数:21
相关论文
共 24 条
[1]   Distributed output regulation of heterogeneous linear multi-agent systems with communication constraints [J].
Abdessameud, Abdelkader ;
Tayebi, Abdelhamid .
AUTOMATICA, 2018, 91 :152-158
[2]   Analyzing and visualizing multiagent rewards in dynamic and stochastic domains [J].
Agogino, Adrian K. ;
Tumer, Kagan .
AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2008, 17 (02) :320-338
[3]  
[Anonymous], C EMP METH NAT LANG
[4]  
[Anonymous], 2018, ARXIV180909332
[5]  
[Anonymous], 1991, Second International Conference on the Simulation and Synthesis of Living Systems (Artificial Life II)
[6]  
[Anonymous], 2017, ARXIV171208266
[7]  
[Anonymous], AAAI FALL S SER ARL
[8]  
[Anonymous], 34 ACM SIGAPP S APPL
[9]  
[Anonymous], THESIS
[10]  
[Anonymous], 2019, ARXIV190108492