Model-Based Reinforcement Learning for Trajectory Tracking of Musculoskeletal Robots

被引:2
作者
Xu, Haoran [1 ]
Fan, Jianyin [1 ]
Wang, Qiang [1 ]
机构
[1] Harbin Inst Technol, Dept Control Sci & Engn, Harbin, Peoples R China
来源
2023 IEEE INTERNATIONAL INSTRUMENTATION AND MEASUREMENT TECHNOLOGY CONFERENCE, I2MTC | 2023年
基金
中国国家自然科学基金;
关键词
model learning; model-based reinforcement learning; pneumatic actuators; ARM DRIVEN; MUSCLE;
D O I
10.1109/I2MTC53148.2023.10175993
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This paper aims to solve the trajectory tracking task of the pneumatic musculoskeletal robot within a model-based reinforcement learning framework. Considering the limited sensors and short lifespan of self-made pneumatic artificial muscles, multi-task Gaussian process regression is employed for micro-data model learning and the learned model is combined with cross entropy method (CEM)-based model predictive control to plan for the optimal action online. To further compensate for the model imperfection and improve the control performance, a proportional derivative controller-like strategy is proposed to guide the search space of the original CEM solver. The effectiveness of our approach is verified on a real musculoskeletal system with one degree of freedom and the results show that only 50 s of interacting with the environment is enough for the robot to learn trajectory tracking skills from scratch.
引用
收藏
页数:6
相关论文
共 23 条
[1]  
Andrikopoulos G, 2017, MED C CONTR AUTOMAT, P241, DOI 10.1109/MED.2017.7984125
[2]  
[Anonymous], 2007, Advances in Neural Information Processing Systems, DOI DOI 10.5555/2981562.2981582
[3]   Control of Musculoskeletal Systems Using Learned Dynamics Models [J].
Buechler, Dieter ;
Calandra, Roberto ;
Schoelkopf, Bernhard ;
Peters, Jan .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04) :3161-3168
[4]  
Büchler D, 2016, IEEE INT CONF ROBOT, P4086, DOI 10.1109/ICRA.2016.7487599
[5]   A Review of Algorithms for Compliant Control of Stiff and Fixed-Compliance Robots [J].
Calanca, Andrea ;
Muradore, Riccardo ;
Fiorini, Paolo .
IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2016, 21 (02) :613-624
[6]  
Chua K, 2018, ADV NEUR IN, V31
[7]  
Daerden Frank, 2002, European Journal of Mechanical and Environmental Engineering, V47, P11
[8]  
Deisenroth M.P., 2010, Efficient reinforcement learning using Gaussian processes
[9]  
Gardner JR, 2018, ADV NEUR IN, V31
[10]  
Geist AR., 2021, GAMM MITTEILUNGEN, V44