Fully Automated Annotation With Noise-Masked Visual Markers for Deep-Learning-Based Object Detection

被引:19
作者
Kiyokawa, Takuya [1 ]
Tomochika, Keita [1 ]
Takamatsu, Jun [1 ]
Ogasawara, Tsukasa [1 ]
机构
[1] Nara Inst Sci & Technol, Div Informat Sci, Nara 6300192, Japan
关键词
Computer vision for automation; deep learning in robotics and automation; object detection; segmentation and categorization; POSE ESTIMATION;
D O I
10.1109/LRA.2019.2899153
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Automated factories use deep-learning-based vision systems to accurately detect various products. However, training such vision systems requires manual annotation of a significant amount of data to optimize the large number of parameters of the deep convolutional neural networks. Such manual annotation is very time-consuming and laborious. To reduce this burden, we propose a fully automated annotation approach without any manual intervention. To do this, we associate one visual marker with one object and capture them in the same image. However, if an image showing the marker is used for training, normally, the neural network learns the marker as a feature of the object. By hiding the marker with a noise mask, we succeeded in reducing this erroneous learning. Experiments verified the effectiveness of the proposed method in comparison with manual annotation, in terms of both the time needed to collect training data and the resulting detection accuracy of the vision system. The time required for data collection was reduced from 16.1 to 1.87 h. The accuracy of the vision system trained with the proposed method was 87.3%, which is higher than the accuracy of a vision system trained with the manual method.
引用
收藏
页码:1972 / 1977
页数:6
相关论文
共 17 条
[1]  
Adhikari B, 2018, EUR W VIS INF PROCES
[2]  
[Anonymous], ARXIV170306370
[3]  
Cubuk Ekin D., 2018, ARXIV180509501
[4]   Automatic generation and detection of highly reliable fiducial markers under occlusion [J].
Garrido-Jurado, S. ;
Munoz-Salinas, R. ;
Madrid-Cuevas, F. J. ;
Marin-Jimenez, M. J. .
PATTERN RECOGNITION, 2014, 47 (06) :2280-2292
[5]   SSD-6D: Making RGB-Based 3D Detection and 6D Pose Estimation Great Again [J].
Kehl, Wadim ;
Manhardt, Fabian ;
Tombari, Federico ;
Ilic, Slobodan ;
Navab, Nassir .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :1530-1538
[6]   SSD: Single Shot MultiBox Detector [J].
Liu, Wei ;
Anguelov, Dragomir ;
Erhan, Dumitru ;
Szegedy, Christian ;
Reed, Scott ;
Fu, Cheng-Yang ;
Berg, Alexander C. .
COMPUTER VISION - ECCV 2016, PT I, 2016, 9905 :21-37
[7]  
Maiettini E, 2017, 2017 IEEE-RAS 17TH INTERNATIONAL CONFERENCE ON HUMANOID ROBOTICS (HUMANOIDS), P862, DOI 10.1109/HUMANOIDS.2017.8246973
[8]  
Montanari D, 2017, ESSCIRC 2017 - 43RD IEEE EUROPEAN SOLID STATE CIRCUITS CONFERENCE, P27, DOI 10.1109/ESSCIRC.2017.8094517
[9]   Learning Deep Object Detectors from 3D Models [J].
Peng, Xingchao ;
Sun, Baochen ;
Ali, Karim ;
Saenko, Kate .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1278-1286
[10]   YouTube-BoundingBoxes: A Large High-Precision Human-Annotated Data Set for Object Detection in Video [J].
Real, Esteban ;
Shlens, Jonathon ;
Mazzocchi, Stefano ;
Pan, Xin ;
Vanhoucke, Vincent .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :7464-7473