Visual Measurement Method and Application of Mobile Manipulator Pose Estimation Based on PPMCC-IMM Filtering

被引:0
|
作者
Zhang, Shijun [1 ]
Cheng, Shuhong [2 ]
Jin, Zhenlin [1 ]
机构
[1] Yanshan Univ, Inst Mech Engn, Qinhuangdao 066004, Hebei, Peoples R China
[2] Yanshan Univ, Inst Elect Engn, Qinhuangdao 066004, Hebei, Peoples R China
基金
中国国家自然科学基金;
关键词
Manipulators; Visualization; Manipulator dynamics; Robots; Pose estimation; Mathematical models; Current measurement; Mobile manipulator; pose estimation; visual measurement; FIDUCIAL MARKERS; ROBOT; CALIBRATION; SYSTEM;
D O I
10.1109/TIM.2023.3268468
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Increasing demand for mobile manipulators in fields requiring high-precision tasks has introduced new requirements regarding their kinematic accuracy. Pose measurement of the mobile manipulator is a crucial method to improve the absolute positioning accuracy of the robot. To solve the pose estimation of the mobile manipulator, we propose a space measurement method based on visual marks and an interactive filtering method to improve the stability of measurement results. First, the kinematics and dynamics of the mobile manipulator are analyzed, and the mathematical expression of the model is obtained to obtain the theoretical trajectory. Second, a visual marker based on pyramid hierarchy is proposed, which can adapt to scenes with significant distance scale changes. A visual marker of a regular 12-face is designed, the marker of each face is uniquely coded to ensure that multiple markers can be captured at the same time under a single viewing angle, and geometric relations determine the distance and pose of the end. We propose a filtering method considering the correlation of historical data to improve the stability and accuracy of measurement data. The matching degree of the filter motion mode is improved through the Pearson coefficient relationship between the filtered data in the current period and the historical data. In this article, several experiments verify the method's effectiveness. The repeatability of the measurement can reach 0.75 mm, and the mean error (MAV) is 0.009 mm. The noise test proves the anti-interference ability of the filtering method. The new filtering method has excellent performance on various tracking tracks. All experiments show that the method proposed in this article has a good measurement effect and is a low-cost, convenient, and efficient method to solve the pose estimation and precise trajectory error analysis of mobile manipulators.
引用
收藏
页数:12
相关论文
共 26 条
  • [21] A method for multi-person mobile virtual reality fire evacuation drills based on pose estimation: Consistency of vision and perception
    Dang, Pei
    Zhu, Jun
    Cao, Yungang
    Wu, Jianlin
    Li, Weilian
    Hu, Ya
    You, Jigang
    Fu, Lin
    SAFETY SCIENCE, 2024, 170
  • [22] Research on pose estimation for stereo vision measurement system by an improved method: uncertainty weighted stereopsis pose solution method based on projection vector
    Cui, Jiashan
    Min, Changwan
    Feng, Dongzhu
    OPTICS EXPRESS, 2020, 28 (04) : 5470 - 5491
  • [23] Research on the Method of Universal Forklift Pallet Detection and Pose Estimation Based on Visual and 3D Point Cloud Fusion
    Gu, Jingxin
    Song, Hongchao
    2024 5TH INTERNATIONAL CONFERENCE ON MECHATRONICS TECHNOLOGY AND INTELLIGENT MANUFACTURING, ICMTIM 2024, 2024, : 388 - 394
  • [24] Enhanced Monocular Visual Odometry: A Robust Appearance-Based Method for Accurate Vehicle Pose Estimation
    Rajesh, R.
    Manivannan, P. V.
    IEEE ACCESS, 2024, 12 : 106176 - 106192
  • [25] Robust and accurate 2D-tracking-based 3D positioning method: Application to head pose estimation
    Ariz, Mikel
    Villanueva, Arantxa
    Cabeza, Rafael
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2019, 180 : 13 - 22
  • [26] Measurement of mobile manipulator chassis pose change caused by suspension deformation and end-effector accuracy improvement based on multi-sensor fusion
    Feng, Yixiao
    Tian, Xiangyu
    Li, Tiemin
    Jiang, Yao
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2023, 170