Steel Roll Eye Pose Detection Based on Binocular Vision and Mask R-CNN

被引:0
作者
Su, Xuwu [1 ]
Wang, Jie [1 ]
Wang, Yifan [1 ]
Zhang, Daode [1 ]
机构
[1] Hubei Univ Technol, Sch Mech Engn, Wuhan 430000, Peoples R China
关键词
pose estimation; binocular vision technology; image segmentation; feature extraction; deep learning;
D O I
10.3390/s25061805
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
To achieve automation at the inner corner guard installation station in a steel coil packaging production line and enable automatic docking and installation of the inner corner guard after eye position detection, this paper proposes a binocular vision method based on deep learning for eye position detection of steel coil rolls. The core of the method involves using the Mask R-CNN algorithm within a deep-learning framework to identify the target region and obtain a mask image of the steel coil end face. Subsequently, the binarized image of the steel coil end face was processed using the RGB vector space image segmentation method. The target feature pixel points were then extracted using Sobel edges, and the parameters were fitted by the least-squares method to obtain the deflection angle and the horizontal and vertical coordinates of the center point in the image coordinate system. Through the ellipse parameter extraction experiment, the maximum deviations in the pixel coordinate system for the center point in the u and v directions were 0.49 and 0.47, respectively. The maximum error in the deflection angle was 0.45 degrees. In the steel coil roll eye position detection experiments, the maximum deviations for the pitch angle, deflection angle, and centroid coordinates were 2.17 degrees, 2.24 degrees, 3.53 mm, 4.05 mm, and 4.67 mm, respectively, all of which met the actual installation requirements. The proposed method demonstrates strong operability in practical applications, and the steel coil end face position solving approach significantly enhances work efficiency, reduces labor costs, and ensures adequate detection accuracy.
引用
收藏
页数:23
相关论文
共 32 条
  • [1] Fused RetinaNet for small target detection in aerial images
    Ahmed, M.
    Wang, Y.
    Maher, Ali
    Bai, X.
    [J]. INTERNATIONAL JOURNAL OF REMOTE SENSING, 2022, 43 (08) : 2813 - 2836
  • [2] An Improved Faster R-CNN for Small Object Detection
    Cao, Changqing
    Wang, Bo
    Zhang, Wenrui
    Zeng, Xiaodong
    Yan, Xu
    Feng, Zhejun
    Liu, Yutao
    Wu, Zengyan
    [J]. IEEE ACCESS, 2019, 7 : 106838 - 106846
  • [3] Integrating explainable AI and depth cameras to achieve automation in grasping Operations: A case study of shoe company
    Chiu, Ming-Chuan
    Yang, Li-Sheng
    [J]. ADVANCED ENGINEERING INFORMATICS, 2024, 62
  • [4] The circular mark projection error compensation in camera calibration
    Cui, Jia-shan
    Huo, Ju
    Yang, Ming
    [J]. OPTIK, 2015, 126 (20): : 2458 - 2463
  • [5] Histograms of oriented gradients for human detection
    Dalal, N
    Triggs, B
    [J]. 2005 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 1, PROCEEDINGS, 2005, : 886 - 893
  • [6] Automatic Control of Food Packaging Machinery
    Dong Tianfei
    [J]. 2020 6TH INTERNATIONAL CONFERENCE ON ADVANCES IN ENERGY, ENVIRONMENT AND CHEMICAL ENGINEERING, PTS 1-5, 2020, 546
  • [7] Detection of Parking Slots Based on Mask R-CNN
    Jiang, Shaokang
    Jiang, Haobin
    Ma, Shidian
    Jiang, Zhongxu
    [J]. APPLIED SCIENCES-BASEL, 2020, 10 (12):
  • [8] A Survey of Deep Learning-Based Object Detection
    Jiao, Licheng
    Zhang, Fan
    Liu, Fang
    Yang, Shuyuan
    Li, Lingling
    Feng, Zhixi
    Qu, Rong
    [J]. IEEE ACCESS, 2019, 7 : 128837 - 128868
  • [9] Surface defect detection of vehicle light guide plates based on an improved RetinaNet
    Li, Junfeng
    Wang, Hao
    [J]. MEASUREMENT SCIENCE AND TECHNOLOGY, 2022, 33 (04)
  • [10] Learning latent geometric consistency for 6D object pose estimation in heavily cluttered scenes
    Li, Qingnan
    Hu, Ruimin
    Xiao, Jing
    Wang, Zhongyuan
    Chen, Yu
    [J]. JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2020, 70