Few-Shot Object Detection Algorithm Based on Geometric Prior and Attention RPN

被引:1
作者
Chen, Xiu [1 ]
Li, Yujie [1 ]
Lu, Huimin [2 ]
机构
[1] Yangzhou Univ, Sch Informat Engn, Yangzhou, Jiangsu, Peoples R China
[2] Southeast Univ, Inst Adv Ocean Res Nantong, Sch Automat, Nanjing, Peoples R China
来源
20TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE, IWCMC 2024 | 2024年
关键词
Few-Shot Object Detection; Geometric Similarity; Synthetic Data Sets; Attention RPN;
D O I
10.1109/IWCMC61514.2024.10592343
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Intelligent factories driven by deep vision technology use robotic arms to perform tasks such as picking and assembling in a production environment. In practical applications, with the continuous changes of products on the industrial pipeline, the detection model needs to continuously train new weights to adapt to new application scenarios. It is time-consuming and labor-intensive to manually collect training data when deploying the production line, and it cannot be quickly adapted in industrial scenarios. Therefore, we propose an attention RPN (Region Proposal Network) few-shot object detection algorithm based on geometric prior. The algorithm uses the attention RPN module to strengthen the feature extraction ability of the detection model and uses the virtual simulation software to generate synthetic data similar to the real object geometry as the base class data to train the feature extraction network so that the network obtains the ability to extract geometric features on the base class object. By comparing the learning strategies, only a small number of real data samples are used to train the detection model twice. The experimental results show that the algorithm can detect more objects than the existing few-shot object detection algorithm in the industrial scene with only a small amount of real sample data, and the detection accuracy can reach 97%.
引用
收藏
页码:706 / 711
页数:6
相关论文
共 36 条
[11]  
Huang G., 2022, IEEE T PATTERN ANAL
[12]   Momentum Contrast for Unsupervised Visual Representation Learning [J].
He, Kaiming ;
Fan, Haoqi ;
Wu, Yuxin ;
Xie, Saining ;
Girshick, Ross .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :9726-9735
[13]   RepMet: Representative-based metric learning for classification and few-shot object detection [J].
Karlinsky, Leonid ;
Shtok, Joseph ;
Harary, Sivan ;
Schwartz, Eli ;
Aides, Amit ;
Feris, Rogerio ;
Giryes, Raja ;
Bronstein, Alex M. .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :5192-5201
[14]  
Khosla P, 2020, ADV NEUR IN, V33
[15]  
Köhler M, 2022, Arxiv, DOI [arXiv:2112.11699, 10.48550/arXiv.2112.11699, DOI 10.48550/ARXIV.2112.11699]
[16]  
Kolve E, 2022, Arxiv, DOI [arXiv:1712.05474, 10.48550/ARXIV.1712.05474]
[17]   Spatio-Temporal Attention Networks for Action Recognition and Detection [J].
Li, Jun ;
Liu, Xianglong ;
Zhang, Wenxuan ;
Zhang, Mingyuan ;
Song, Jingkuan ;
Sebe, Nicu .
IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (11) :2990-3001
[18]  
Li Y., 2022, IEEE Intell. Syst.
[19]   Real-time optical flow-based video stabilization for unmanned aerial vehicles [J].
Lim, Anli ;
Ramesh, Bharath ;
Yang, Yue ;
Xiang, Cheng ;
Gao, Zhi ;
Lin, Feng .
JOURNAL OF REAL-TIME IMAGE PROCESSING, 2019, 16 (06) :1975-1985
[20]   A digital twin-based sim-to-real transfer for deep reinforcement learning-enabled industrial robot grasping [J].
Liu, Yongkui ;
Xu, He ;
Liu, Ding ;
Wang, Lihui .
ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING, 2022, 78