DIGITAL TWIN SIMULATIONS BASED REINFORCEMENT LEARNING FOR NAVIGATION AND CONTROL OF A WHEEL-ON-LEG MOBILE ROBOT

被引:0
作者
Alsaleh, Saleh [1 ]
Tepljakov, Aleksei [1 ]
Tamre, Mart [2 ]
Kuts, Vladimir [3 ]
Petlenkov, Eduard [1 ]
机构
[1] Tallinn Univ Technol, Dept Comp Syst, EE-19086 Tallinn, Harjumaa, Estonia
[2] Tallinn Univ Technol, Sch Engn, EE-19086 Tallinn, Harjumaa, Estonia
[3] Tallinn Univ Technol, Mech & Ind Engn Dept, EE-19086 Tallinn, Harjumaa, Estonia
来源
PROCEEDINGS OF ASME 2022 INTERNATIONAL MECHANICAL ENGINEERING CONGRESS AND EXPOSITION, IMECE2022, VOL 2B | 2022年
关键词
Reinforcement learning; digital twins; mobile robots; machine learning; simulation; LEVEL;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Hybrid mobile robots are able to function in a number of different modes of locomotion, which increases their capacity to overcome challenges and makes them appropriate for a wide range of applications. To be able to develop navigation techniques that make use of these improved capabilities, one must first have a solid grasp of the constraints imposed by each of those different modalities of locomotion. In this paper, we present a data-driven approach for evaluating the robots' locomotion modes. To do this, we formalize the problem as a reinforcement learning task that is applied to a digital twin simulation of the mobile robot. The proposed method is demonstrated through the use of a case study that examines the capabilities of hybrid wheel-on-leg robot locomotion modes in terms of speed, slope ascent, and step obstacle climbing. First, a comprehensive explanation of the process of creating the digital twin of the mobile robot through the use of the Unity gaming engine is presented. Second, a description of the construction of three test environments is provided so that the aforementioned capabilities of the robot can be evaluated. In the end, Reinforcement Learning is used to evaluate the two types of locomotion that the mobile robot can utilize in each of these different environments. Corresponding simulations are conducted in the virtual environment and the results are analyzed.
引用
收藏
页数:8
相关论文
共 21 条
[1]   Towards Artificial Intelligence Driven Immersive Environments in Virtual Reality for Industrial Applications [J].
Alsaleh, Saleh ;
Tepljakov, Aleksei ;
Tamre, Mart ;
Petlenkov, Eduard .
2021 44TH INTERNATIONAL CONFERENCE ON TELECOMMUNICATIONS AND SIGNAL PROCESSING (TSP), 2021, :340-345
[2]   An optional passive/active transformable wheel-legged mobility concept for search and rescue robots [J].
Bai, Long ;
Guan, Jian ;
Chen, Xiaohong ;
Hou, Junzhan ;
Duan, Wenbo .
ROBOTICS AND AUTONOMOUS SYSTEMS, 2018, 107 :145-155
[3]   Reinforcement Learning Approach for Optimal Distributed Energy Management in a Microgrid [J].
Foruzan, Elham ;
Soh, Leen-Kiat ;
Asgarpoor, Sohrab .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2018, 33 (05) :5749-5758
[4]   Autonomous Navigation for Quadrupedal Robots with Optimized Jumping through Constrained Obstacles [J].
Gilroy, Scott ;
Lau, Derek ;
Yang, Lizhi ;
Izaguirre, Ed ;
Biermayer, Kristen ;
Xiao, Anxing ;
Sun, Mengti ;
Agrawal, Ayush ;
Zeng, Jun ;
Li, Zhongyu ;
Sreenath, Koushil .
2021 IEEE 17TH INTERNATIONAL CONFERENCE ON AUTOMATION SCIENCE AND ENGINEERING (CASE), 2021, :2132-2139
[5]  
Hiiemaa M., 2013, Motion Planner for Skid-Steer Unmanned Ground Vehicle
[6]  
Juliani A, 2020, Arxiv, DOI arXiv:1809.02627
[7]  
Malayjerdi M., 2021, Virtual Simulations Environment Development for Autonomous Vehicles Interaction
[8]  
Marchesini Enrico, 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA), P10688, DOI 10.1109/ICRA40945.2020.9196739
[9]  
Mnih V, 2013, Arxiv, DOI arXiv:1312.5602
[10]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533