Knowledge Reuse of Multi-Agent Reinforcement Learning in Cooperative Tasks

被引:2
作者
Shi, Daming [1 ]
Tong, Junbo [1 ]
Liu, Yi [1 ]
Fan, Wenhui [1 ]
机构
[1] Tsinghua Univ, Dept Automat, Beijing 100084, Peoples R China
关键词
multi-agent; reinforcement learning; cooperative task; adding teammate; knowledge sharing; knowledge transferring;
D O I
10.3390/e24040470
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
With the development and appliance of multi-agent systems, multi-agent cooperation is becoming an important problem in artificial intelligence. Multi-agent reinforcement learning (MARL) is one of the most effective methods for solving multi-agent cooperative tasks. However, the huge sample complexity of traditional reinforcement learning methods results in two kinds of training waste in MARL for cooperative tasks: all homogeneous agents are trained independently and repetitively, and multi-agent systems need training from scratch when adding a new teammate. To tackle these two problems, we propose the knowledge reuse methods of MARL. On the one hand, this paper proposes sharing experience and policy within agents to mitigate training waste. On the other hand, this paper proposes reusing the policies learned by original teams to avoid knowledge waste when adding a new agent. Experimentally, the Pursuit task demonstrates how sharing experience and policy can accelerate the training speed and enhance the performance simultaneously. Additionally, transferring the learned policies from the N-agent enables the (N+1)-agent team to immediately perform cooperative tasks successfully, and only a minor training resource can allow the multi-agents to reach optimal performance identical to that from scratch.
引用
收藏
页数:15
相关论文
共 31 条
[1]  
[Anonymous], 1986, BCSG201028 BOEING AD
[2]  
[Anonymous], 2013, P 2013 INT C AUT AG
[3]  
Barrett S., 2011, P INT C AUT AG MULT
[4]   A comprehensive survey of multiagent reinforcement learning [J].
Busoniu, Lucian ;
Babuska, Robert ;
De Schutter, Bart .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2008, 38 (02) :156-172
[5]   A Review of the Applications of Agent Technology in Traffic and Transportation Systems [J].
Chen, Bo ;
Cheng, Harry H. .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2010, 11 (02) :485-497
[6]  
da Silva FL, 2017, AAMAS'17: PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, P1100
[7]   Enabling scalable and fault-tolerant multi-agent systems by utilizing cloud-native computing [J].
Daehling, Stefan ;
Razik, Lukas ;
Monti, Antonello .
AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2021, 35 (01)
[8]   A Cooperative Spectrum Sensing With Multi-Agent Reinforcement Learning Approach in Cognitive Radio Networks [J].
Gao, Ang ;
Du, Chengyuan ;
Ng, Soon Xin ;
Liang, Wei .
IEEE COMMUNICATIONS LETTERS, 2021, 25 (08) :2604-2608
[9]  
Glatt R, 2016, PROCEEDINGS OF 2016 5TH BRAZILIAN CONFERENCE ON INTELLIGENT SYSTEMS (BRACIS 2016), P91, DOI [10.1109/BRACIS.2016.17, 10.1109/BRACIS.2016.027]
[10]   Model Learning and Knowledge Sharing for Cooperative Multiagent Systems in Stochastic Environment [J].
Jiang, Wei-Cheng ;
Narayanan, Vignesh ;
Li, Jr-Shin .
IEEE TRANSACTIONS ON CYBERNETICS, 2021, 51 (12) :5717-5727