Object pose estimation method for robotic arm grasping

被引:0
作者
Huang C. [1 ]
Hou S. [1 ]
机构
[1] Department of Automation, Harbin University of Science and Technology, Harbin University of Science and Technology, Heilongjiang Province, Harbin City
关键词
attention mechanism; convolutional neural network; feature fusion; planar grasp; Pose estimation;
D O I
10.3233/JIFS-234351
中图分类号
学科分类号
摘要
To address the issue of target detection in the planar grasping task, a position and attitude estimation method based on YOLO-Pose is proposed. The aim is to detect the three-dimensional position of the spacecraft’s center point and the planar two-dimensional attitude in real time. First, the weight is trained through transfer learning, and the number of key points is optimized by analyzing the shape characteristics of the spacecraft to improve the representation of pose information. Second, the CBAM dual-channel attention mechanism is integrated into the C3 module of the backbone network to improve the accuracy of pose estimation. Furthermore, the Wing Loss function is used to mitigate the problem of random offset in key points. The incorporation of the bi-directional feature pyramid network (BiFPN) structure into the neck network further improves the accuracy of target detection. The experimental results show that the average accuracy value of the optimized algorithm has increased. The average detection speed can meet the speed and accuracy requirements of the actual capture task and has practical application value. © 2024 – IOS Press. All rights reserved.
引用
收藏
页码:10787 / 10803
页数:16
相关论文
共 18 条
  • [1] Lu H., Research on Plane Grasp Detection Method Based on RGB-D Images
  • [2] Xia J., Qian K., Ma X.-D., Et al., Rapid detection of robotic planar grasp poses based on cascaded convolutional neural networks, Robotics, 40, 6, pp. 794-802, (2018)
  • [3] Lin S.-X., Fang Q.-S., Contour registration method for workpieces based on improved SIFT algorithm, Mechanical Engineer, 10, pp. 4-7, (2022)
  • [4] Chan R., Tam P., A new Hough transform based position estimation algorithm, Proceedings of ANZIIS’94 – Australian New Zealand Intelligent Information Systems Conference, pp. 140-144, (1994)
  • [5] Song C., Song J., Huang Q., HybridPose: 6D Object Pose Estimation under Hybrid Representations, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, 2020, pp. 428-437
  • [6] Xiang Y., Schmidt T., Narayanan V., Et al., PoseCNN: A Convolutional Neural Network for 6D Object Pose Estimation in Cluttered Scenes, (2017)
  • [7] Zhou X., Lan X., Zhang H., Et al., Fully Convolutional Grasp Detection Network with Oriented Anchor Box, Proceedings of the IEEE Conference on Intelligent Robots and Systems (IROS), pp. 7223-7230, (2018)
  • [8] He K., Zhang X., Ren S., Et al., Deep Residual Learning for Image Recognition, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770-778, (2016)
  • [9] Peng S., Zhou Y., Liu Y., Et al., PVNet: Pixel-wise Voting Network for 6DoF Pose Estimation, IEEE Transactions on Pattern Analysis and Machine Intelligence, pp. 3212-3223, (2022)
  • [10] Qi C.-R., Wei L., Wu C., Et al., Frustum PointNets for 3D Object Detection from RGB-D Data, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 918-927, (2018)