Object-aware interactive perception for tabletop scene exploration

被引:1
作者
Koc, Cagatay [1 ]
Sariel, Sanem [1 ]
机构
[1] Istanbul Tech Univ, Fac Comp & Informat Engn, Artificial Intelligence & Robot Lab, Maslak, Turkiye
关键词
Cognitive robotics; Active perception; Interactive perception; View evaluation; Robotic scene exploration; Volumetric utility metrics; ACTIVE VISION; POSE ESTIMATION; MANIPULATION; RECOGNITION; RECONSTRUCTION; SYSTEMS;
D O I
10.1016/j.robot.2024.104674
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recent advancements in sensors and deep learning techniques have improved reliability of robotic perceptual systems, but current systems are not robust enough for real -world challenges such as occlusions and sensing uncertainties in cluttered scenes. To overcome these issues, active or interactive perception actions are often necessary, such as sensor repositioning or object manipulation to reveal more information about the scene. Existing perception systems lack a comprehensive approach that incorporates both active and interactive action spaces, thereby limiting the robot's perception capabilities. Moreover, these systems focus on exploring a single object or scene, without utilizing object information to guide the exploration of multiple objects. In this work, we propose an object -aware hybrid perception system that selects the next best action by considering both active and interactive action spaces and enhances the selection process with an object -aware approach to guide the cognitive robot operating in tabletop scenarios. Novel volumetric utility metrics are used to evaluate actions that include positioning sensors from a heterogeneous set or manipulating objects to gain a better perspective of the scene. The proposed system maintains the volumetric information of the scene that includes semantic information about objects, enabling it to exploit object information, associate occlusion with corresponding objects, and make informed decisions about object manipulation. We evaluate the performance of our system both in simulated and real -world experiments using a Baxter robotic platform equipped with two arms, RGB and depth cameras. Our experimental results show that the proposed system outperforms the compared state-of-the-art methods in the given scenarios, achieving an 11.2% performance increase.
引用
收藏
页数:15
相关论文
共 70 条
[1]  
Almeida D, 2019, IEEE-RAS INT C HUMAN, P297, DOI [10.1109/humanoids43949.2019.9035001, 10.1109/Humanoids43949.2019.9035001]
[2]  
Atanasov N, 2013, IEEE INT CONF ROBOT, P4216, DOI 10.1109/ICRA.2013.6631173
[3]   Revisiting active perception [J].
Bajesy, Ruzena ;
Aloimonos, Yiannis ;
Tsotsos, John K. .
AUTONOMOUS ROBOTS, 2018, 42 (02) :177-196
[4]   Interactive Perception: Leveraging Action in Perception and Perception in Action [J].
Bohg, Jeannette ;
Hausman, Karol ;
Sankaran, Bharath ;
Brock, Oliver ;
Kragic, Danica ;
Schaal, Stefan ;
Sukhatme, Gaurav S. .
IEEE TRANSACTIONS ON ROBOTICS, 2017, 33 (06) :1273-1291
[5]   Closed-Loop Next-Best-View Planning for Target-Driven Grasping [J].
Breyer, Michel ;
Ott, Lionel ;
Siegwart, Roland ;
Chung, Jen Jen .
2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, :1411-1416
[6]  
Browatzki B, 2012, IEEE INT CONF ROBOT, P2021, DOI 10.1109/ICRA.2012.6225218
[7]  
Burusa AK, 2024, Arxiv, DOI [arXiv:2306.09801, 10.1016/j.biosystemseng.2024.09.018, DOI 10.48550/ARXIV.2306.09801, DOI 10.1016/J.BIOSYSTEMSENG.2024.09.018]
[8]  
Burusa AK, 2024, Arxiv, DOI arXiv:2206.10274
[9]   Active Vision via Extremum Seeking for Robots in Unstructured Environments: Applications in Object Recognition and Manipulation [J].
Calli, Berk ;
Caarls, Wouter ;
Wisse, Martijn ;
Jonker, Pieter P. .
IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2018, 15 (04) :1810-1822
[10]   Benchmarking in Manipulation Research Using the Yale-CMU-Berkeley Object and Model Set [J].
Calli, Berk ;
Walsman, Aaron ;
Singh, Arjun ;
Srinivasa, Siddhartha ;
Abbeel, Pieter ;
Dollar, Aaron M. .
IEEE ROBOTICS & AUTOMATION MAGAZINE, 2015, 22 (03) :36-52