A YOLO-GGCNN based grasping framework for mobile robots in unknown environments

被引:20
作者
Li, Zhen [1 ]
Xu, Benlian [1 ]
Wu, Di [2 ]
Zhao, Kang [3 ]
Chen, Siwen [4 ]
Lu, Mingli [5 ]
Cong, Jinliang [5 ]
机构
[1] Suzhou Univ Sci & Technol, Sch Elect & informat Engn, Suzhou 215009, Jiangsu, Peoples R China
[2] Northwestern Polytech Univ, Sch Automat, Xian 710072, Shaanxi, Peoples R China
[3] Changshu Inst Technol, Sch Mech Engn, Suzhou 215500, Jiangsu, Peoples R China
[4] Nanjing Univ, Software Inst, Nanjing 210009, Jiangsu, Peoples R China
[5] Changshu Inst Technol, Sch Elect & Automat Engn, Suzhou 215500, Jiangsu, Peoples R China
关键词
Optimization scheduling; Grasping system; Visual recognition; Object localization; FEATURES; SLAM; DEEP;
D O I
10.1016/j.eswa.2023.119993
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To grasp a desirable object in an unknown environment, both accurate map construction and reliable visual recognition are prerequisites for the collaboration of mobile robots. We propose a multirobot cooperative simultaneous localization and mapping (SLAM)-based map fusion algorithm which is developed to construct a map of an unknown environment through multiple exploring robots with high maneuverability. The resulting map is then transmitted to the mobile robotic arm through the Robot Operating System (ROS) for the upcoming navigation step. In the grasping module, a two-step cascaded system, i.e., YOLOv4 and a generative grasping convolutional neural network (YOLO-GGCNN), is proposed to grasp any given object via the mobile robotic arm. The capture accuracy of our algorithm is 86.0%, and the detection time required for a single frame is 0.11 s. In the recognition module, the grasping prediction is finally converted into an information flow that controls the operation of the robotic arm. The proposed framework is validated through applications in various environments and comparisons with existing approaches. The results indicate that our approach allows the mobile robotic arm to selectively and accurately grasp objects of interest in unknown environments, and it is more reliable than other methods.
引用
收藏
页数:14
相关论文
共 56 条
  • [1] FastSLAM 2.0 tracking and mapping as a Cloud Robotics service
    Ali, Shimaa S.
    Hammad, Abdallah
    Eldien, Adly S. Tag
    [J]. COMPUTERS & ELECTRICAL ENGINEERING, 2018, 69 : 412 - 421
  • [2] Comprehensive review on brain-controlled mobile robots and robotic arms based on electroencephalography signals
    Aljalal, Majid
    Ibrahim, Sutrisno
    Djemal, Ridha
    Ko, Wonsuk
    [J]. INTELLIGENT SERVICE ROBOTICS, 2020, 13 (04) : 539 - 563
  • [3] Reconfigurable Pavement Sweeping Robot and Pedestrian Cohabitant Framework by Vision Techniques
    Anh Vu Le
    Hayat, Abdullah Aamir
    Elara, Mohan Rajesh
    Nguyen Huu Khanh Nhan
    Prathap, Kandasamy
    [J]. IEEE ACCESS, 2019, 7 : 159402 - 159414
  • [4] Automated Object Manipulation Using Vision-Based Mobile Robotic System for Construction Applications
    Asadi, Khashayar
    Haritsa, Varun R.
    Han, Kevin
    Ore, John-Paul
    [J]. JOURNAL OF COMPUTING IN CIVIL ENGINEERING, 2021, 35 (01)
  • [5] Simultaneous localization and mapping (SLAM): Part II
    Bailey, Tim
    Durrant-Whyte, Hugh
    [J]. IEEE ROBOTICS & AUTOMATION MAGAZINE, 2006, 13 (03) : 108 - 117
  • [6] Barrau A, 2016, Arxiv, DOI arXiv:1510.06263
  • [7] Bay H., 2006, EUROPEAN C COMPUTER, P404
  • [8] Bochkovskiy A, 2020, Arxiv, DOI arXiv:2004.10934
  • [9] Data-Driven Grasp Synthesis-A Survey
    Bohg, Jeannette
    Morales, Antonio
    Asfour, Tamim
    Kragic, Danica
    [J]. IEEE TRANSACTIONS ON ROBOTICS, 2014, 30 (02) : 289 - 309
  • [10] Caldera Shehan, 2018, Multimodal Technologies and Interaction, V2, DOI 10.3390/mti2030057