Predictive Visuo-Tactile Interactive Perception Framework for Object Properties Inference

被引:0
作者
Dutta, Anirvan [1 ,2 ]
Burdet, Etienne [2 ]
Kaboli, Mohsen [1 ,3 ]
机构
[1] RoboTac Lab, BMW Grp, D-85748 Munich, Germany
[2] Imperial Coll Sci, Technol & Med, London SW7 2AZ, England
[3] Eindhoven Univ Technol, NL-5612 AZ Eindhoven, Netherlands
关键词
Robots; Robot sensing systems; Shape; Sensors; Friction; Visualization; Uncertainty; Point cloud compression; Bayes methods; Probabilistic logic; Active interactive perception; recursive Bayesian filtering; visual and tactile sensing; PARAMETERS; NETWORKS; VISION;
D O I
10.1109/TRO.2025.3531816
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Interactive exploration of unknown objects' properties, such as stiffness, mass, center of mass, friction coefficient, and shape, is crucial for autonomous robotic systems operating in unstructured environments. Precise identification of these properties is essential for stable and controlled object manipulation and for anticipating the outcomes of (prehensile or nonprehensile) manipulation actions, such as pushing, pulling, and lifting. Our study focuses on autonomously inferring the physical properties of a diverse set of homogeneous, heterogeneous, and articulated objects using a robotic system equipped with vision and tactile sensors. We propose a novel predictive perception framework to identify object properties by leveraging versatile exploratory actions: nonprehensile pushing and prehensile pulling. A key component of our framework is a novel active shape perception mechanism that seamlessly initiates exploration. In addition, our dual differentiable filtering with graph neural networks learns the object-robot interaction and enables consistent inference of indirectly observable, time-invariant object properties. Finally, we develop a N-step information gain approach to select the most informative actions for efficient learning and inference. Extensive real-robot experiments with planar objects show that our predictive perception framework outperforms state-of-the-art baselines and showcases it in three major applications for object tracking, goal-driven task, and environmental change detection.
引用
收藏
页码:1386 / 1403
页数:18
相关论文
共 84 条
[1]  
Agrawal P, 2016, ADV NEUR IN, V29
[2]  
ALOIMONOS J, 1987, INT J COMPUT VISION, V1, P333
[3]  
[Anonymous], [61] [Online]. Available: https://www.osengines.com/.%20[Accessed%20May%202018].
[4]  
[Anonymous], OptiTrack Motion Capture Systems
[5]   LEAST-SQUARES FITTING OF 2 3-D POINT SETS [J].
ARUN, KS ;
HUANG, TS ;
BLOSTEIN, SD .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1987, 9 (05) :699-700
[6]   ESTIMATION OF INERTIAL PARAMETERS OF MANIPULATOR LOADS AND LINKS [J].
ATKESON, CG ;
AN, CH ;
HOLLERBACH, JM .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 1986, 5 (03) :101-119
[7]   Revisiting active perception [J].
Bajesy, Ruzena ;
Aloimonos, Yiannis ;
Tsotsos, John K. .
AUTONOMOUS ROBOTS, 2018, 42 (02) :177-196
[8]   CONDITIONS FOR VERSATILE LEARNING, HELMHOLTZ UNCONSCIOUS INFERENCE, AND THE TASK OF PERCEPTION [J].
BARLOW, H .
VISION RESEARCH, 1990, 30 (11) :1561-1571
[9]  
Barr A. H., 1981, IEEE Computer Graphics and Applications, V1, P11, DOI 10.1109/MCG.1981.1673799
[10]  
Bohg Jeannette, 2011, IEEE International Conference on Robotics and Automation, P686