Learning what to memorize: Using intrinsic motivation to form useful memory in partially observable reinforcement learning

被引:1
|
作者
Demir, Alper [1 ]
机构
[1] Izmir Univ Econ, Dept Comp Engn, TR-35330 Izmir, Turkiye
关键词
Memory; Intrinsic motivation; Partial observability; Reinforcement learning; AGENTS;
D O I
10.1007/s10489-022-04328-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement Learning faces an important challenge in partially observable environments with long-term dependencies. In order to learn in an ambiguous environment, an agent has to keep previous perceptions in a memory. Earlier memory-based approaches use a fixed method to determine what to keep in the memory, which limits them to certain problems. In this study, we follow the idea of giving the control of the memory to the agent by allowing it to take memory-changing actions. Thus, the agent becomes more adaptive to the dynamics of an environment. Further, we formalize an intrinsic motivation to support this learning mechanism, which guides the agent to memorize distinctive events and enable it to disambiguate its state in the environment. Our overall approach is tested and analyzed on several partial observable tasks with long-term dependencies. The experiments show a clear improvement in terms of learning performance compared to other memory based methods.
引用
收藏
页码:19074 / 19092
页数:19
相关论文
共 50 条
  • [1] Learning what to memorize: Using intrinsic motivation to form useful memory in partially observable reinforcement learning
    Alper Demir
    Applied Intelligence, 2023, 53 : 19074 - 19092
  • [2] Learning reward machines: A study in partially observable reinforcement learning 
    Icarte, Rodrigo Toro
    Klassen, Toryn Q.
    Valenzano, Richard
    Castro, Margarita P.
    Waldie, Ethan
    Mcilraith, Sheila A.
    ARTIFICIAL INTELLIGENCE, 2023, 323
  • [3] Intrinsic Motivation and Introspection in Reinforcement Learning
    Merrick, Kathryn E.
    IEEE TRANSACTIONS ON AUTONOMOUS MENTAL DEVELOPMENT, 2012, 4 (04) : 315 - 329
  • [4] Mnemonic Dictionary Learning for Intrinsic Motivation in Reinforcement Learning
    Yan, Renye
    Wu, Zhe
    Zhan, Yuan
    Tao, Pin
    Wang, Zongwei
    Cai, Yimao
    Xing, Junliang
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [5] Using Emotions as Intrinsic Motivation to Accelerate Classic Reinforcement Learning
    Lu, Cheng-Xiang
    Sun, Zhi-Yuan
    Shi, Zhong-Zhi
    Cao, Bao-Xiang
    2016 INTERNATIONAL CONFERENCE ON INFORMATION SYSTEM AND ARTIFICIAL INTELLIGENCE (ISAI 2016), 2016, : 332 - 337
  • [6] Feature Control as Intrinsic Motivation for Hierarchical Reinforcement Learning
    Dilokthanakul, Nat
    Kaplanis, Christos
    Pawlowski, Nick
    Shanahan, Murray
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (11) : 3409 - 3418
  • [7] Benefits of Combining Dimensional Attention and Working Memory for Partially Observable Reinforcement Learning Problems
    Omatu, Ngozi
    Phillips, Joshua L.
    ACMSE 2021: PROCEEDINGS OF THE 2021 ACM SOUTHEAST CONFERENCE, 2021, : 209 - 213
  • [8] Partially Observable Reinforcement Learning for Sustainable Active Surveillance
    Chen, Hechang
    Yang, Bo
    Liu, Jiming
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2018, PT II, 2018, 11062 : 425 - 437
  • [9] Modeling and reinforcement learning in partially observable many-agent systems
    He, Keyang
    Doshi, Prashant
    Banerjee, Bikramjit
    AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2024, 38 (01)
  • [10] Adaptive Compensation for Robotic Joint Failures Using Partially Observable Reinforcement Learning
    Pham, Tan-Hanh
    Aikins, Godwyll
    Truong, Tri
    Nguyen, Kim-Doang
    ALGORITHMS, 2024, 17 (10)