Fast conflict resolution based on reinforcement learning in multi-agent system

被引:0
作者
Piao, SH [1 ]
Hong, BR [1 ]
Chu, HT [1 ]
机构
[1] Harbin Inst Technol, Sch Comp Sci & Technol, Harbin 150001, Peoples R China
来源
CHINESE JOURNAL OF ELECTRONICS | 2004年 / 13卷 / 01期
关键词
cooperation; reinforcement learning; multi-agent system; conflict resolution;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In multi-agent system where each agent has a different goal (even the team of agents has the same goal), agents must be able to resolve conflicts arising in the process of achieving their goal. Many researchers presented methods for conflict resolution, e.g., Reinforcement learning (RL), but the conventional RL requires a large computation cost because every agent must learn, at the same time the overlap of actions selected by each agent results in local conflict. Therefore in this paper, we propose a novel method to solve these problems. In order to deal with the conflict within the multi-agent system, the concept of potential field function based Action selection priority level (ASPL) is brought forward. In this method, all kinds of environment factor that may have influence on the priority are effectively computed with the potential field function. So the priority to access the local resource can be decided rapidly. By avoiding the complex coordination mechanism used in general multi-agent system, the conflict in multi-agent system is settled more efficiently. Our system consists of RL with ASPL module and generalized rules module. Using ASPL, RL module chooses a proper cooperative behavior, and generalized rule module can accelerate the learning process. By applying the proposed method to Robot Soccer, the learning process can be accelerated. The results of simulation and real experiments indicate the effectiveness of the method.
引用
收藏
页码:92 / 95
页数:4
相关论文
共 12 条
[1]  
BLICHEV G, 1996, P IEEE INT C EV COMP, P832
[2]  
CAITAO C, 2000, P IEEE INT C INT ROB, P1397
[3]  
Desai JP, 1998, IEEE INT CONF ROBOT, P2864, DOI 10.1109/ROBOT.1998.680621
[4]  
Inoue K, 2000, 2000 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2000), VOLS 1-3, PROCEEDINGS, P885, DOI 10.1109/IROS.2000.893131
[5]  
Kawakami K, 2001, ISIE 2001: IEEE INTERNATIONAL SYMPOSIUM ON INDUSTRIAL ELECTRONICS PROCEEDINGS, VOLS I-III, P423, DOI 10.1109/ISIE.2001.931827
[6]  
Kitano H, 1997, AI MAG, V18, P73
[7]   Cooperation protocols in multi-agent robotic systems [J].
Lin, FC ;
Hsu, JYJ .
AUTONOMOUS ROBOTS, 1997, 4 (02) :175-198
[8]  
LIN LJ, 1992, P 2 INT C SIM AD BEH, V2, P271
[9]  
LITTMAN ML, 1994, P 11 INT C MACH LEAR, P157
[10]   Reinforcement learning in the multi-robot domain [J].
Mataric, MJ .
AUTONOMOUS ROBOTS, 1997, 4 (01) :73-83