The next-best-view for workpiece localization in robot workspace

被引:2
作者
Hu, Jie [1 ]
Pagilla, Prabhakar R. [1 ]
Darbha, Swaroop [1 ]
机构
[1] Texas A&M Univ, Dept Mech Engn, College Stn, TX 77843 USA
来源
2021 IEEE/ASME INTERNATIONAL CONFERENCE ON ADVANCED INTELLIGENT MECHATRONICS (AIM) | 2021年
关键词
workpiece localization; robotics; manufacturing; next-best-view;
D O I
10.1109/AIM46487.2021.9517657
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Workpiece localization is the process of obtaining the location of a workpiece in a reference frame of a robotic workspace. The location (position and orientation) is represented by the transformation between a local frame associated with the workpiece and the specified reference frame in the workspace. In this work, we study the workpiece localization problem without the two commonly adopted restrictive assumptions: the data used to calculate the transformation is readily available and the correspondence between the data sets used for calculation is known. The goal is to automate the localization process starting from efficient data collection to determining the workpiece location in the workspace. We describe a strategy that includes the following aspects: predicting the correspondence between the measured data and the workpiece CAD model data; generating representative vectors that would aid in determining the next-best-view for collecting new information of the workpiece location; evaluating a search region to find the next sensor location that satisfies both the robot kinematics as well as sensor field-of-view constraints while giving the maximum view gain; and calculating the rigid body transformation from the local frame to the world frame to localize the workpiece. Numerical simulation and experimental results are presented and discussed for the proposed strategy.
引用
收藏
页码:1201 / 1206
页数:6
相关论文
共 38 条
[31]   The surface edge explorer (SEE): A measurement-direct approach to next best view planning [J].
Border, Rowan ;
Gammell, Jonathan D. .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2024, 43 (10) :1506-1532
[32]   Next Best View Planning via Reinforcement Learning for Scanning of Arbitrary 3D Shapes [J].
Potapova, S. G. ;
Artemov, A. V. ;
Sviridov, S. V. ;
Musatkina, D. A. ;
Zorin, D. N. ;
Burnaev, E. V. .
JOURNAL OF COMMUNICATIONS TECHNOLOGY AND ELECTRONICS, 2020, 65 (12) :1484-1490
[33]   Tree-based search of the next best view/state for three-dimensional object reconstruction [J].
Irving Vasquez-Gomez, J. ;
Enrique Sucar, L. ;
Murrieta-Cid, Rafael ;
Herrera-Lozada, Juan-Carlos .
INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2018, 15 (01)
[34]   Next Best View Planning via Reinforcement Learning for Scanning of Arbitrary 3D Shapes [J].
S. G. Potapova ;
A. V. Artemov ;
S. V. Sviridov ;
D. A. Musatkina ;
D. N. Zorin ;
E. V. Burnaev .
Journal of Communications Technology and Electronics, 2020, 65 :1484-1490
[35]   A next best view method based on self-occlusion information in depth images for moving object [J].
Zhang, Shihui ;
Li, Xin ;
He, Huan ;
Miao, Yuxia .
MULTIMEDIA TOOLS AND APPLICATIONS, 2018, 77 (08) :9753-9777
[36]   A next best view method based on self-occlusion information in depth images for moving object [J].
Shihui Zhang ;
Xin Li ;
Huan He ;
Yuxia Miao .
Multimedia Tools and Applications, 2018, 77 :9753-9777
[37]   Next best view planning for building a 3D model of objects from color images [J].
Zarn, Kristian ;
Skocaj, Danijel .
ELEKTROTEHNISKI VESTNIK, 2020, 87 (03) :75-86
[38]   Next Best View Planning via Deep Reinforcement Leaning for 3D Reconstruction of Turbine Blades [J].
Xiao, Xiang ;
Fang, Qiu ;
Peng, Weixing ;
Wang, Yaonan ;
Hong, Lu ;
Ye, Jun .
PROCEEDINGS OF THE 36TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC 2024, 2024, :4022-4027