Learning Intelligent Behavior in a Non-stationary and Partially Observable Environment

被引:0
|
作者
SelÇuk şenkul
Faruk Polat
机构
[1] Middle East Technical University,Computer Engineering Department
来源
Artificial Intelligence Review | 2002年 / 18卷
关键词
agent learning; multi-agent systems; Q-learning; reinforcement learning;
D O I
暂无
中图分类号
学科分类号
摘要
Individual learning in an environment where more than one agent exist is a chal-lengingtask. In this paper, a single learning agent situated in an environment where multipleagents exist is modeled based on reinforcement learning. The environment is non-stationaryand partially accessible from an agents' point of view. Therefore, learning activities of anagent is influenced by actions of other cooperative or competitive agents in the environment.A prey-hunter capture game that has the above characteristics is defined and experimentedto simulate the learning process of individual agents. Experimental results show that thereare no strict rules for reinforcement learning. We suggest two new methods to improve theperformance of agents. These methods decrease the number of states while keeping as muchstate as necessary.
引用
收藏
页码:97 / 115
页数:18
相关论文
共 50 条
  • [31] Reinforcement learning and evolutionary algorithms for non-stationary multi-armed bandit problems
    Koulouriotis, D. E.
    Xanthopoulos, A.
    APPLIED MATHEMATICS AND COMPUTATION, 2008, 196 (02) : 913 - 922
  • [32] Accelerated Variant of Reinforcement Learning Algorithms for Light Control with Non-stationary User Behaviour
    Haddam, Nassim
    Boulakia, Benjamin Cohen
    Barth, Dominique
    PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON SMART CITIES AND GREEN ICT SYSTEMS (SMARTGREENS), 2022, : 78 - 85
  • [33] Learn to coordinate with generic non-stationary opponents
    Zhang, Kaifu
    PROCEEDINGS OF THE FIFTH IEEE INTERNATIONAL CONFERENCE ON COGNITIVE INFORMATICS, VOLS 1 AND 2, 2006, : 558 - 565
  • [34] A novel approach for self-driving car in partially observable environment using life long reinforcement learning
    Quadir, Md Abdul
    Jaiswal, Dibyanshu
    Mohan, Senthilkumar
    Innab, Nisreen
    Sulaiman, Riza
    Alaoui, Mohammed Kbiri
    Ahmadian, Ali
    SUSTAINABLE ENERGY GRIDS & NETWORKS, 2024, 38
  • [35] Learning optimal admission control in partially observable queueing networks
    Anselmi, Jonatha
    Gaujal, Bruno
    Rebuffi, Louis-Sebastien
    QUEUEING SYSTEMS, 2024, 108 (1-2) : 31 - 79
  • [36] Online Monitoring of Heterogeneous Partially Observable Data Streams Based on Q-Learning
    Li, Haoqian
    Ye, Honghan
    Cheng, Jing-Ru C.
    Liu, Kaibo
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024, : 4802 - 4817
  • [37] ROBOTIC OBSTACLE AVOIDANCE IN A PARTIALLY OBSERVABLE ENVIRONMENT USING FEATURE RANKING
    Gharbieh, Waseem
    Al-Mousa, Amjed
    INTERNATIONAL JOURNAL OF ROBOTICS & AUTOMATION, 2019, 34 (05) : 572 - 579
  • [38] Bayesian Nonparametric Methods for Partially-Observable Reinforcement Learning
    Doshi-Velez, Finale
    Pfau, David
    Wood, Frank
    Roy, Nicholas
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2015, 37 (02) : 394 - 407
  • [39] Reinforcement Learning of Chaotic Systems Control in Partially Observable Environments
    Weissenbacher, Max
    Borovykh, Anastasia
    Rigas, Georgios
    FLOW TURBULENCE AND COMBUSTION, 2025,
  • [40] A robust policy bootstrapping algorithm for multi-objective reinforcement learning in non-stationary environments
    Abdelfattah, Sherif
    Kasmarik, Kathryn
    Hu, Jiankun
    ADAPTIVE BEHAVIOR, 2020, 28 (04) : 273 - 292