Graph and dynamics interpretation in robotic reinforcement learning task

被引:6
|
作者
Yao, Zonggui [1 ]
Yu, Jun [1 ]
Zhang, Jian [2 ]
He, Wei [3 ]
机构
[1] Hangzhou Dianzi Univ, Sch Comp Sci & Technol, Key Lab Complex Syst Modeling & Simulat, Hangzhou 310018, Peoples R China
[2] Zhejiang Int Studies Univ, Sch Sci & Technol, Hangzhou 310012, Peoples R China
[3] Univ Sci & Technol Beijing, Sch Automat & Elect Engn, Beijing 100083, Peoples R China
基金
中国国家自然科学基金;
关键词
Graph neural networks (GNNs); Dynamics estimations; Robotic controls; Robotic force transmission; Trajectory following; Reinforcement learning; GRADIENT;
D O I
10.1016/j.ins.2022.08.041
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Robot control tasks are typically solved by reinforcement learning approaches in a circular way of trial and learn. A recent trend of the research on robotic reinforcement learning is the employment of the deep learning methods. Existing deep learning methods achieve the control by training the approximation models of the dynamic function, value function or the policy function in the control algorithms. However, these methods usually handle the modeling from a statistical perspective without considering the physics characteristics of the robot's motion. One of the typical problems is the force transmission through different parts of a robot and the calculations of the robotic dynamics quantities are prone to be ignored. To this end, we propose to use the force transmission graph to interpret the force transmission mechanism obeyed by the motion of the robot and estimate the dynamics quantities of the robot's motion with a quadratic model. Following this concern, we propose a model-based reinforcement learning framework for robotic control in which the dynamic model comprises two components, i.e. the Graph Convolution Network (GCN) and the Two-Layer Perception (TLP) network. The GCN serves as a parameter estimator of the force transmission graph and a structural feature extractor. The TLP network approximates the quadratic model that should be able to estimate the dynamics quantities of the robot's motion. For this reason, the proposed framework is named as GCN of Dynamics estimation in Reinforcement Learning method (GDRL for short). The deployed method interprets the intrinsic mechanism of robotic force transmissions through robot limbs, therefore the model is highly interpretable. Experimental results show that GDRL can predict the gesture and location of the robot for the next move well such that the performance of our method surpasses that of the previous methods in robot control task in our task setting of multiple types of robots. We also designed to compare with the previous model-free methods in our task setting, and the results are outstanding, which is owing to the interpretation of the physical characteristics. (C) 2022 Elsevier Inc. All rights reserved.
引用
收藏
页码:317 / 334
页数:18
相关论文
共 50 条
  • [21] Safe Reinforcement Learning in a Simulated Robotic Arm
    Kovac, Luka
    Farkas, Igor
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT I, 2023, 14254 : 585 - 589
  • [22] Reinforcement learning in robotic applications: a comprehensive survey
    Bharat Singh
    Rajesh Kumar
    Vinay Pratap Singh
    Artificial Intelligence Review, 2022, 55 : 945 - 990
  • [23] Decentralized reinforcement learning control of a robotic manipulator
    Busoniu, Lucian
    De Schutter, Bart
    Babuska, Robert
    2006 9TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION, VOLS 1- 5, 2006, : 1121 - +
  • [24] Application of Reinforcement Learning to a Robotic Drinking Assistant
    Shastha, Tejas Kumar
    Kyrarini, Maria
    Graeser, Axel
    ROBOTICS, 2020, 9 (01)
  • [25] Learning Continuous Control Actions for Robotic Grasping with Reinforcement Learning
    Shahid, Asad Ali
    Roveda, Loris
    Piga, Dario
    Braghin, Francesco
    2020 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2020, : 4066 - 4072
  • [26] Integration of Evolutionary Computing and Reinforcement Learning for Robotic Imitation Learning
    Tan, Huan
    Balajee, Kannan
    Lynn, DeRose
    2014 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), 2014, : 407 - 412
  • [27] The dynamics of generalized reinforcement learning
    Lahkar, Ratul
    Seymour, Robert M.
    JOURNAL OF ECONOMIC THEORY, 2014, 151 : 584 - 595
  • [28] Utilizing Reinforcement Learning and Causal Graph Networks to Address the Intricate Dynamics in Financial Risk Prediction
    Ma, Fake
    Li, Huwei
    Ilyas, Muhammad
    INTERNATIONAL JOURNAL OF INFORMATION TECHNOLOGIES AND SYSTEMS APPROACH, 2024, 17 (01)
  • [29] Causal Reinforcement Learning for Knowledge Graph Reasoning
    Li, Dezhi
    Lu, Yunjun
    Wu, Jianping
    Zhou, Wenlu
    Zeng, Guangjun
    APPLIED SCIENCES-BASEL, 2024, 14 (06):
  • [30] Survey of Knowledge Graph Based on Reinforcement Learning
    Ma A.
    Yu Y.
    Yang S.
    Shi C.
    Li J.
    Cai X.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2022, 59 (08): : 1694 - 1722