Learning intelligent behavior in a non-stationary and partially observable environment

被引:2
作者
Senkul, S [1 ]
Polat, F [1 ]
机构
[1] Middle E Tech Univ, Dept Comp Engn, TR-06531 Ankara, Turkey
关键词
agent learning; multi-agent systems; Q-learning; reinforcement learning;
D O I
10.1023/A:1019935502139
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Individual learning in an environment where more than one agent exist is a challenging task. In this paper, a single learning agent situated in an environment where multiple agents exist is modeled based on reinforcement learning. The environment is non-stationary and partially accessible from an agents' point of view. Therefore, learning activities of an agent is influenced by actions of other cooperative or competitive agents in the environment. A prey-hunter capture game that has the above characteristics is defined and experimented to simulate the learning process of individual agents. Experimental results show that there are no strict rules for reinforcement learning. We suggest two new methods to improve the performance of agents. These methods decrease the number of states while keeping as much state as necessary.
引用
收藏
页码:97 / 115
页数:19
相关论文
共 28 条
[1]   Multiagent reinforcement learning using function approximation [J].
Abul, O ;
Polat, F ;
Alhajj, R .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2000, 30 (04) :485-497
[2]  
[Anonymous], COMPUTERS THOUGHT
[3]  
[Anonymous], READINGS AGENTS
[4]  
Bellmann R., 1957, DYNAMIC PROGRAMMING
[5]  
ELLIS HC, 1972, FUNDAMENTALS HUMAN L
[6]  
Estes W. K., 1970, Learning theory and mental development
[7]  
Howard R., 1960, DYNAMIC PROGRAMMING
[8]  
HU J, 1998, P 15 INT C MACH LEAR, P242
[9]  
HU J, 1998, GAMES EC BEHAV
[10]  
HULSE SH, 1984, PSYCHOL LEARNING