Further Advantages of Data Augmentation on Convolutional Neural Networks

被引:81
作者
Hernandez-Garcia, Alex [1 ]
Koenig, Peter [1 ]
机构
[1] Univ Osnabruck, Inst Cognit Sci, Osnabruck, Germany
来源
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2018, PT I | 2018年 / 11139卷
关键词
Data augmentation; Regularization; CNNs;
D O I
10.1007/978-3-030-01418-6_10
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data augmentation is a popular technique largely used to enhance the training of convolutional neural networks. Although many of its benefits are well known by deep learning researchers and practitioners, its implicit regularization effects, as compared to popular explicit regularization techniques, such as weight decay and dropout, remain largely unstudied. As a matter of fact, convolutional neural networks for image object classification are typically trained with both data augmentation and explicit regularization, assuming the benefits of all techniques are complementary. In this paper, we systematically analyze these techniques through ablation studies of different network architectures trained with different amounts of training data. Our results unveil a largely ignored advantage of data augmentation: networks trained with just data augmentation more easily adapt to different architectures and amount of training data, as opposed to weight decay and dropout, which require specific fine-tuning of their hyperparameters.
引用
收藏
页码:95 / 103
页数:9
相关论文
共 34 条
[1]  
ABADI M., 2015, TensorFlow: large-scale machine learning on heterogeneous systems
[2]  
[Anonymous], 2013, P MACHINE LEARNING R
[3]   Deep, Big, Simple Neural Nets for Handwritten Digit Recognition [J].
Ciresan, Dan Claudiu ;
Meier, Ueli ;
Gambardella, Luca Maria ;
Schmidhuber, Juergen .
NEURAL COMPUTATION, 2010, 22 (12) :3207-3220
[4]  
[Anonymous], 2015, Deep image: Scaling up image recognition
[5]  
[Anonymous], 2011, AISTATS
[6]  
Antoniou A, 2017, CoRR
[7]  
Bengio Y., 2006, ADV NEURAL INFORM PR, V19
[8]  
Chollet F., 2015, about us
[9]  
DeVries T, 2017, INT C LEARN REPR
[10]  
DeVries T., 2017, arXiv