Graph and dynamics interpretation in robotic reinforcement learning task

被引:6
|
作者
Yao, Zonggui [1 ]
Yu, Jun [1 ]
Zhang, Jian [2 ]
He, Wei [3 ]
机构
[1] Hangzhou Dianzi Univ, Sch Comp Sci & Technol, Key Lab Complex Syst Modeling & Simulat, Hangzhou 310018, Peoples R China
[2] Zhejiang Int Studies Univ, Sch Sci & Technol, Hangzhou 310012, Peoples R China
[3] Univ Sci & Technol Beijing, Sch Automat & Elect Engn, Beijing 100083, Peoples R China
基金
中国国家自然科学基金;
关键词
Graph neural networks (GNNs); Dynamics estimations; Robotic controls; Robotic force transmission; Trajectory following; Reinforcement learning; GRADIENT;
D O I
10.1016/j.ins.2022.08.041
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Robot control tasks are typically solved by reinforcement learning approaches in a circular way of trial and learn. A recent trend of the research on robotic reinforcement learning is the employment of the deep learning methods. Existing deep learning methods achieve the control by training the approximation models of the dynamic function, value function or the policy function in the control algorithms. However, these methods usually handle the modeling from a statistical perspective without considering the physics characteristics of the robot's motion. One of the typical problems is the force transmission through different parts of a robot and the calculations of the robotic dynamics quantities are prone to be ignored. To this end, we propose to use the force transmission graph to interpret the force transmission mechanism obeyed by the motion of the robot and estimate the dynamics quantities of the robot's motion with a quadratic model. Following this concern, we propose a model-based reinforcement learning framework for robotic control in which the dynamic model comprises two components, i.e. the Graph Convolution Network (GCN) and the Two-Layer Perception (TLP) network. The GCN serves as a parameter estimator of the force transmission graph and a structural feature extractor. The TLP network approximates the quadratic model that should be able to estimate the dynamics quantities of the robot's motion. For this reason, the proposed framework is named as GCN of Dynamics estimation in Reinforcement Learning method (GDRL for short). The deployed method interprets the intrinsic mechanism of robotic force transmissions through robot limbs, therefore the model is highly interpretable. Experimental results show that GDRL can predict the gesture and location of the robot for the next move well such that the performance of our method surpasses that of the previous methods in robot control task in our task setting of multiple types of robots. We also designed to compare with the previous model-free methods in our task setting, and the results are outstanding, which is owing to the interpretation of the physical characteristics. (C) 2022 Elsevier Inc. All rights reserved.
引用
收藏
页码:317 / 334
页数:18
相关论文
共 50 条
  • [1] Evaluating Task Optimization and Reinforcement Learning Models in Robotic Task Parameterization
    Delledonne, Michele
    Villagrossi, Enrico
    Beschi, Manuel
    Rastegarpanah, Alireza
    IEEE ACCESS, 2024, 12 : 173734 - 173748
  • [2] The Use of Reinforcement Learning in the Task of Moving Objects with the Robotic Arm
    Aitygulov, Ermek E.
    ARTIFICIAL INTELLIGENCE, 2019, 11866 : 119 - 126
  • [3] Robotic Disassembly Task Training and Skill Transfer Using Reinforcement Learning
    Qu, Mo
    Wang, Yongjing
    Pham, Duc Truong
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2023, 19 (11) : 10934 - 10943
  • [4] Influence Graph based Task Decomposition and State Abstraction in Reinforcement Learning
    Yu, Lasheng
    Hong, Fei
    Wang, PengRen
    Xu, Yang
    Liu, Yong
    PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE FOR YOUNG COMPUTER SCIENTISTS, VOLS 1-5, 2008, : 136 - 141
  • [5] Control strategy of robotic manipulator based on multi-task reinforcement learning
    Wang, Tao
    Ruan, Ziming
    Wang, Yuyan
    Chen, Chong
    COMPLEX & INTELLIGENT SYSTEMS, 2025, 11 (03)
  • [6] Federated Deep Reinforcement Learning for Task Scheduling in Heterogeneous Autonomous Robotic System
    Tai Manh Ho
    Kim-Khoa Nguyen
    Cheriet, Mohamed
    2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 1134 - 1139
  • [7] Efficient Reinforcement Learning of Task Planners for Robotic Palletization Through Iterative Action Masking Learning
    Wu, Zheng
    Li, Yichuan
    Zhan, Wei
    Liu, Changliu
    Liu, Yun-Hui
    Tomizuka, Masayoshi
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (11): : 9303 - 9310
  • [8] Task graph offloading via deep reinforcement learning in mobile edge computing
    Liu, Jiagang
    Mi, Yun
    Zhang, Xinyu
    Li, Xiaocui
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 158 : 545 - 555
  • [9] Intelligent Task Allocation for Mobile Crowdsensing With Graph Attention Network and Deep Reinforcement Learning
    Xu, Chenghao
    Song, Wei
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2023, 10 (02): : 1032 - 1048
  • [10] Graph Representations for Reinforcement Learning
    Schab, Esteban
    Casanova, Carlos
    Piccoli, Fabiana
    JOURNAL OF COMPUTER SCIENCE & TECHNOLOGY, 2024, 24 (01): : 29 - 38