Geometric Reinforcement Learning for Robotic Manipulation

被引:5
作者
Alhousani, Naseem [1 ,3 ]
Saveriano, Matteo [2 ,4 ]
Sevinc, Ibrahim [3 ]
Abdulkuddus, Talha [2 ]
Kose, Hatice [1 ]
Abu-Dakka, Fares J. [5 ]
机构
[1] Istanbul Tech Univ, Fac Comp & Informat Engn, Sariyer, TR-80333 Maslak, Istanbul, Turkiye
[2] ILITRON Enerji Bilgi Teknolojileri AŞ, Kagıthane, TR-34415 Istanbul, Turkiye
[3] MCFLY Robot Teknolojileri AŞ, Sarıyer, TR-34485 Istanbul, Turkiye
[4] Univ Trento, Dept Ind Engn DII, I-38123 Trento, Italy
[5] Tech Univ Munich, Munich Inst Robot & Machine Intelligence MIRMI, D-80992 Munich, Germany
关键词
Learning on manifolds; policy optimization; policy search; geometric reinforcement learning; RIEMANNIAN-MANIFOLDS;
D O I
10.1109/ACCESS.2023.3322654
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning (RL) is a popular technique that allows an agent to learn by trial and error while interacting with a dynamic environment. The traditional Reinforcement Learning (RL) approach has been successful in learning and predicting Euclidean robotic manipulation skills such as positions, velocities, and forces. However, in robotics, it is common to encounter non-Euclidean data such as orientation or stiffness, and failing to account for their geometric nature can negatively impact learning accuracy and performance. In this paper, to address this challenge, we propose a novel framework for RL that leverages Riemannian geometry, which we call Geometric Reinforcement Learning (G-RL), to enable agents to learn robotic manipulation skills with non-Euclidean data. Specifically, G-RL utilizes the tangent space in two ways: a tangent space for parameterization and a local tangent space for mapping to a non-Euclidean manifold. The policy is learned in the parameterization tangent space, which remains constant throughout the training. The policy is then transferred to the local tangent space via parallel transport and projected onto the non-Euclidean manifold. The local tangent space changes over time to remain within the neighborhood of the current manifold point, reducing the approximation error. Therefore, by introducing a geometrically grounded pre- and post-processing step into the traditional RL pipeline, our G-RL framework enables several model-free algorithms designed for Euclidean space to learn from non-Euclidean data without modifications. Experimental results, obtained both in simulation and on a real robot, support our hypothesis that G-RL is more accurate and converges to a better solution than approximating non-Euclidean data.
引用
收藏
页码:111492 / 111505
页数:14
相关论文
共 44 条
[1]  
Absil PA, 2008, OPTIMIZATION ALGORITHMS ON MATRIX MANIFOLDS, P1
[2]   A probabilistic framework for learning geometry-based robot manipulation skills [J].
Abu-Dakka, Fares J. ;
Huang, Yanlong ;
Silverio, Joao ;
Kyrki, Ville .
ROBOTICS AND AUTONOMOUS SYSTEMS, 2021, 141 (141)
[3]  
Abu-Dakka FJ, 2020, IEEE INT CONF ROBOT, P4421, DOI [10.1109/icra40945.2020.9196952, 10.1109/ICRA40945.2020.9196952]
[4]  
Abu-Dakka FJ, 2018, IEEE-RAS INT C HUMAN, P278
[5]  
Beik-Mohammadi H, 2021, Arxiv, DOI [arXiv:2106.04315, 10.48550/arXiv.2106.04315]
[6]   Geometric Deep Learning Going beyond Euclidean data [J].
Bronstein, Michael M. ;
Bruna, Joan ;
LeCun, Yann ;
Szlam, Arthur ;
Vandergheynst, Pierre .
IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (04) :18-42
[7]   Gaussians on Riemannian Manifolds: Applications for Robot Learning and Adaptive Control [J].
Calinon, Sylvain .
IEEE ROBOTICS & AUTOMATION MAGAZINE, 2020, 27 (02) :33-45
[8]   Impedance Adaptation by Reinforcement Learning with Contact Dynamic Movement Primitives [J].
Chang, Chunyang ;
Haninger, Kevin ;
Shi, Yunlei ;
Yuan, Chengjie ;
Chen, Zhaopeng ;
Zhang, Jianwei .
2022 IEEE/ASME INTERNATIONAL CONFERENCE ON ADVANCED INTELLIGENT MECHATRONICS (AIM), 2022, :1185-1191
[9]  
Chatzilygeroudis K, 2017, IEEE INT C INT ROBOT, P51, DOI 10.1109/IROS.2017.8202137
[10]  
Chen J., 2022, P IEEE CVF C COMP VI, P6646