Transfer learning with Partially Constrained Models: Application to reinforcement learning of linked multicomponent robot system control

被引:17
|
作者
Fernandez-Gauna, Borja [1 ]
Manuel Lopez-Guede, Jose [1 ]
Grana, Manuel [1 ]
机构
[1] Univ Basque Country, Computat Intelligence Grp, UPV EHU, Bilbao, Spain
关键词
Reinforcement learning; Linked multicomponent robotic systems; Transfer learning; Hose transportation; MDPS;
D O I
10.1016/j.robot.2012.07.020
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Transfer learning is a hierarchical approach to reinforcement learning of complex tasks modeled as Markov Decision Processes. The learning results on the source task are used as the starting point for the learning on the target task. In this paper we deal with a hierarchy of constrained systems, where the source task is an under-constrained system, hence called the Partially Constrained Model (PCM). Constraints in the framework of reinforcement learning are dealt with by state-action veto policies. We propose a theoretical background for the hierarchy of training refinements, showing that the effective action repertoires learnt on the PCM are maximal, and that the PCM-optimal policy gives maximal state value functions. We apply the approach to learn the control of Linked Multicomponent Robotic Systems using Reinforcement Learning. The paradigmatic example is the transportation of a hose. The system has strong physical constraints and a large state space. Learning experiments in the target task are realized over an accurate but computationally expensive simulation of the hose dynamics. The PCM is obtained simplifying the hose model. Learning results of the PCM Transfer Learning show an spectacular improvement over conventional Q-learning on the target task. (C) 2012 Elsevier B.V. All rights reserved.
引用
收藏
页码:694 / 703
页数:10
相关论文
共 50 条
  • [1] Reinforcement Learning endowed with safe veto policies to learn the control of Linked-Multicomponent Robotic Systems
    Fernandez-Gauna, Borja
    Grana, Manuel
    Manuel Lopez-Guede, Jose
    Etxeberria-Agiriano, Ismael
    Ansoategui, Igor
    INFORMATION SCIENCES, 2015, 317 : 25 - 47
  • [2] Reinforcement learning for robot control
    Smart, WD
    Kaelbling, LP
    MOBILE ROBOTS XVI, 2002, 4573 : 92 - 103
  • [3] Learning for a Robot: Deep Reinforcement Learning, Imitation Learning, Transfer Learning
    Hua, Jiang
    Zeng, Liangcai
    Li, Gongfa
    Ju, Zhaojie
    SENSORS, 2021, 21 (04) : 1 - 21
  • [4] APPLICATION OF REINFORCEMENT LEARNING TO A TWO DOF ROBOT ARM CONTROL
    Albers, Albert
    Yan Wenjie
    Frietsch, Markus
    ANNALS OF DAAAM FOR 2009 & PROCEEDINGS OF THE 20TH INTERNATIONAL DAAAM SYMPOSIUM, 2009, 20 : 415 - 416
  • [5] Application of reinforcement learning in robot soccer
    Duan, Yong
    Liu, Qiang
    Xu, Xinhe
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2007, 20 (07) : 936 - 950
  • [6] Efficient Reinforcement Learning for StarCraft by Abstract Forward Models and Transfer Learning
    Liu, Ruo-Ze
    Guo, Haifeng
    Ji, Xiaozhong
    Yu, Yang
    Pang, Zhen-Jia
    Xiao, Zitai
    Wu, Yuzhou
    Lu, Tong
    IEEE TRANSACTIONS ON GAMES, 2022, 14 (02) : 294 - 307
  • [7] Robot Reinforcement Learning Based on Learning Classifier System
    Shao, Jie
    Yang, Jing-yu
    ADVANCED INTELLIGENT COMPUTING THEORIES AND APPLICATIONS, 2010, 93 : 200 - 207
  • [8] Transfer and Reinforcement Learning Based Production Control
    Steinbacher L.
    Pering E.
    Freitag M.
    ZWF Zeitschrift fuer Wirtschaftlichen Fabrikbetrieb, 2022, 117 (09): : 609 - 613
  • [9] Reinforcement learning from expert demonstrations with application to redundant robot control
    Ramirez, Jorge
    Yu, Wen
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 119
  • [10] Adaptive Control of a Mechatronic System Using Constrained Residual Reinforcement Learning
    Staessens, Tom
    Lefebvre, Tom
    Crevecoeur, Guillaume
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2022, 69 (10) : 10447 - 10456