GAN-Based Data Augmentation with Vehicle Color Changes to Train a Vehicle Detection CNN

被引:6
作者
Ayub, Aroona [1 ]
Kim, HyungWon [2 ]
机构
[1] Chungbuk Natl Univ, Grad Sch Comp Sci, Cheongju 28644, South Korea
[2] Chungbuk Natl Univ, Coll Elect & Comp Engn, Cheongju 28644, South Korea
基金
新加坡国家研究基金会;
关键词
convolutional neural network (CNN); generative adversarial network (GAN); data augmentation; generative models; object detection models;
D O I
10.3390/electronics13071231
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Object detection is a challenging task that requires a lot of labeled data to train convolutional neural networks (CNNs) that can achieve human-level accuracy. However, such data are not easy to obtain, as they involve significant manual work and costs to annotate the objects in images. Researchers have used traditional data augmentation techniques to increase the amount of training data available to them. A recent trend in object detection is to use generative models to automatically create annotated data that can enrich a training set and improve the performance of the target model. This paper presents a method of training the proposed ColorGAN network, which is used to generate augmented data for the target domain of interest with the least compromise in quality. We demonstrate a method to train a GAN with images of vehicles in different colors. Then, we demonstrate that our ColorGAN can change the color of vehicles of any given vehicle dataset to a set of specified colors, which can serve as an augmented training dataset. Our experimental results show that the augmented dataset generated by the proposed method helps enhance the detection performance of a CNN for applications where the original training data are limited. Our experiments also show that the model can achieve a higher mAP of 76% when the model is trained with augmented images along with the original training dataset.
引用
收藏
页数:14
相关论文
共 17 条
[1]  
Ayub A., 2022, P 8 INT C NEXT GEN C, P261
[2]  
Ayub Aroona, 2023, [JOURNAL OF KOREA MULTIMEDIA SOCIETY, 멀티미디어학회논문지], V26, P713
[3]   YOLACT Real-time Instance Segmentation [J].
Bolya, Daniel ;
Zhou, Chong ;
Xiao, Fanyi ;
Lee, Yong Jae .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :9156-9165
[4]  
Choi Y, 2020, Arxiv, DOI [arXiv:1912.01865, 10.48550/arXiv.1912.01865]
[5]   StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation [J].
Choi, Yunjey ;
Choi, Minje ;
Kim, Munyoung ;
Ha, Jung-Woo ;
Kim, Sunghun ;
Choo, Jaegul .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8789-8797
[6]  
Horev R., Lyrn.AI
[7]   A Style-Based Generator Architecture for Generative Adversarial Networks [J].
Karras, Tero ;
Laine, Samuli ;
Aila, Timo .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4396-4405
[8]  
Li XL, 2021, Arxiv, DOI arXiv:2103.05422
[9]   Microsoft COCO: Common Objects in Context [J].
Lin, Tsung-Yi ;
Maire, Michael ;
Belongie, Serge ;
Hays, James ;
Perona, Pietro ;
Ramanan, Deva ;
Dollar, Piotr ;
Zitnick, C. Lawrence .
COMPUTER VISION - ECCV 2014, PT V, 2014, 8693 :740-755
[10]   A Deep Learning-Based Approach to Progressive Vehicle Re-identification for Urban Surveillance [J].
Liu, Xinchen ;
Liu, Wu ;
Mei, Tao ;
Ma, Huadong .
COMPUTER VISION - ECCV 2016, PT II, 2016, 9906 :869-884