GAN-based data augmentation for semantic segmentation in multiple weathers

被引:0
作者
Nakashima K.
Satoh Y.
Kataoka H.
机构
来源
Seimitsu Kogaku Kaishi/Journal of the Japan Society for Precision Engineering | 2021年 / 87卷 / 01期
关键词
Data augmentation; GAN; Multiple weathers; Semantic segmentation; Traffic scene;
D O I
10.2493/jjspe.87.107
中图分类号
学科分类号
摘要
Datasets play an important role in determining the features that deep neural networks can acquire, but they can also contain unintended biases when constructing datasets. The BDD100K dataset, famous for its semantic segmentation task, was collected to include traffic scenes for multiple weather conditions. However, due to differences in frequency of occurrence, there is a bias in the number of data for each weather condition. Therefore, the segmentation network trained by BDD100K has poor recognition performance in some weather conditions. Semantic segmentation is an urgent issue because it is expected to be applied to traffic scene recognition systems. In this paper, we aim to improve the performance of semantic segmentation by designing a method that generates images of desired weather conditions and uses them for data augmentation. In our experiments, we first show that the image generation method we have developed produces images of a quality that can be used for data augmentation. Next, we examine the effect of data augmentation on the semantic segmentation task. As a result, compared to baseline, the mean intersection over union (mloU) improved by about 15% in wet weather, about 9% at night, and about 1% overall. © 2021 Japan Society for Precision Engineering. All rights reserved.
引用
收藏
页码:107 / 113
页数:6
相关论文
共 19 条
  • [1] Deng I., Dong W., Socher R., Li L. -J., Li K., Fei-Fei L., Imagenet: A large-scale hierarchical image database, IEEE Conference on Computer Vision and Pattern Recognition, (2009)
  • [2] Rahman S., Khan S., Porikli F., Zero-shot object detection: Learning to simultaneously recognize and localize novel concepts, Asian Conference on Computer Vision, (2018)
  • [3] Yu F., Chen H., Wang X., Xian W., Chen Y., Liu F., Madhavan, Darrell T., BDDIOOK: A diverse driving dataset for heterogeneous multitask learning, IEEE Conference on Computer Vision and Pattern Recognition, (2020)
  • [4] Lafferty J., McCallum A., Pereira F., Conditional random fields: Probabilistic models for segmenting and labeling sequence data, International Conference on Machine Learning, (2001)
  • [5] Shotton J., Winn J., Rother C., Criininisi A., Textonboost for image understanding: Multi-class object recognition and segmentation by jointly modeling texture, layout, and context, International Journal of Computer Vision, 81, (2009)
  • [6] Long J., Shelhamer E., Darrell T., Fully convohittonal networks for semantic segmentation, IEEE Conference on Computer Vision and Pattern Recognition, (2015)
  • [7] Badrinarayanan V., Kendall A., Cipolla R., Segnet: A deep convolutional encoder-decoder architecture for image segmentation, (2015)
  • [8] Ronncbcrgcr O., Fischer P., Brox T., (J-net: Convolutional networks for biomedical image segmentation, International Conference on Medical Image Computing and Computer-Assisted Intervention, (2015)
  • [9] Chen L. -C., Papandrcou G., Schroff F., Adam H., Rethinking atrous convolution for semantic image segmentation, (2017)
  • [10] Chen L. -C., Papandreou G., Kokkinos I., Murphy K., Yuille A. L., DeepLab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected errs, IEEE Transactions on Pattern Analysis and Machine Intelligence, 40, 4, (2017)