Implementation of Reinforcement Learning by Transfering Sub-Goal Policies in Robot Navigation

被引:0
|
作者
Gokce, Baris [1 ]
Akin, H. Levent [1 ]
机构
[1] Bogazici Univ, Bilgisayar Muhendisligi Bolumu, Istanbul, Turkey
来源
2013 21ST SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU) | 2013年
关键词
Reinforcement Learning; Hierarchical Reinforcement Learning; Transfer Learning; Robot Navigation; SKILL ACQUISITION;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Although Reinforcement Learning (RL) is one of the most popular learning methods, it suffers from the curse of dimensionality. If the state and action domains of the problem are immense, the learning rate of the agent decreases dramatically and eventually the agent loses the ability to learn. In order to eliminate the effects of the curse of the dimensionality, researchers typically concentrate on the methods that reduce the complexity of the problems. While some of them model the problem in a hierarchical manner, the others try to transfer the knowledge obtained during the learning process of simpler tasks. While learning from scratch ignores the previous experiences, transferring full knowledge may mislead the agent because of the conflicting requirements. The main goal of this study is to improve the learning rate of the agent by transferring the relevant parts of the knowledge acquired as a result of previous experiences. The main contribution of this study is to merge these two approaches to transfer only the relevant knowledge in a setting. The proposed method is tested on a robot navigation task in a simulated room-based environment.
引用
收藏
页数:4
相关论文
共 50 条
  • [31] Modular deep reinforcement learning from reward and punishment for robot navigation
    Wang, Jiexin
    Elfwing, Stefan
    Uchibe, Eiji
    NEURAL NETWORKS, 2021, 135 : 115 - 126
  • [32] Local and soft feature selection for value function approximation in batch reinforcement learning for robot navigation
    Fatemeh Fathinezhad
    Peyman Adibi
    Bijan Shoushtarian
    Jocelyn Chanussot
    The Journal of Supercomputing, 2024, 80 : 10720 - 10745
  • [33] Local and soft feature selection for value function approximation in batch reinforcement learning for robot navigation
    Fathinezhad, Fatemeh
    Adibi, Peyman
    Shoushtarian, Bijan
    Chanussot, Jocelyn
    JOURNAL OF SUPERCOMPUTING, 2024, 80 (08) : 10720 - 10745
  • [34] Graph Relational Reinforcement Learning for Mobile Robot Navigation in Large-Scale Crowded Environments
    Liu, Zhe
    Zhai, Yu
    Li, Jiaming
    Wang, Guangming
    Miao, Yanzi
    Wang, Hesheng
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (08) : 8776 - 8787
  • [35] Predictive reinforcement learning: map-less navigation method for mobile robot
    Dobriborsci, Dmitrii
    Zashchitin, Roman
    Kakanov, Mikhail
    Aumer, Wolfgang
    Osinenko, Pavel
    JOURNAL OF INTELLIGENT MANUFACTURING, 2023, 35 (8) : 4217 - 4232
  • [36] RESEARCH ON AUTONOMOUS NAVIGATION AND CONTROL ALGORITHM OF INTELLIGENT ROBOT BASED ON REINFORCEMENT LEARNING
    Yi, Yunlong
    Guan, Ying
    SCALABLE COMPUTING-PRACTICE AND EXPERIENCE, 2025, 26 (01): : 423 - 431
  • [37] Multi mobile robot navigation using distributed value function reinforcement learning
    Babvey, S
    Momtahan, O
    Meybodi, MR
    2003 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1-3, PROCEEDINGS, 2003, : 957 - 962
  • [38] Adaptive state construction for reinforcement learning and its application to robot navigation problems
    Handa, H
    Ninomiya, A
    Horiuchi, T
    Konishi, T
    Baba, M
    2001 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS, VOLS 1-5: E-SYSTEMS AND E-MAN FOR CYBERNETICS IN CYBERSPACE, 2002, : 1436 - 1441
  • [39] RL-DOVS: Reinforcement Learning for Autonomous Robot Navigation in Dynamic Environments
    Mackay, Andrew K.
    Riazuelo, Luis
    Montano, Luis
    SENSORS, 2022, 22 (10)
  • [40] Efficient Reinforcement Learning for 3D LiDAR Navigation of Mobile Robot
    Zhai, Yu
    Liu, Zhe
    Miao, Yanzi
    Wang, Hesheng
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 3755 - 3760