Data Augmentation and Random Multi-Model Deep Learning for Data Classification

被引:0
作者
Harby F. [1 ]
Thaljaoui A. [1 ]
Nayab D. [2 ]
Aladhadh S. [3 ]
Khediri S.E.L. [3 ,4 ]
Khan R.U. [3 ]
机构
[1] Computer Science Department, Future Academy-Higher Future Institute for Specialized Technological Studies
[2] Department of Computer Systems Engineering, Faculty of Electrical and Computer Engineering, University of Engineering and Technology, Peshawar
[3] Department of Information Technology, College of Computer, Qassim University, Buraydah
[4] Department of Computer Sciences, Faculty of Sciences of Gafsa, University of Gafsa, Gafsa
关键词
classification; Data augmentation; generative adversarial networks; machine learning; random multi-model deep learning;
D O I
10.32604/cmc.2022.029420
中图分类号
学科分类号
摘要
In the machine learning (ML) paradigm, data augmentation serves as a regularization approach for creating ML models. The increase in the diversification of training samples increases the generalization capabilities, which enhances the prediction performance of classifiers when tested on unseen examples. Deep learning (DL) models have a lot of parameters, and they frequently overfit. Effectively, to avoid overfitting, data plays a major role to augment the latest improvements in DL. Nevertheless, reliable data collection is a major limiting factor. Frequently, this problem is undertaken by combining augmentation of data, transfer learning, dropout, and methods of normalization in batches. In this paper, we introduce the application of data augmentation in the field of image classification using Random Multi-model Deep Learning (RMDL) which uses the association approaches of multi- DL to yield random models for classification. We present a methodology for using Generative Adversarial Networks (GANs) to generate images for data augmenting. Through experiments, we discover that samples generated by GANs when fed into RMDL improve both accuracy and model efficiency. Experimenting across both MNIST and CIAFAR-10 datasets show that, error rate with proposed approach has been decreased with different random models. © 2023 Authors. All rights reserved.
引用
收藏
页码:5191 / 5207
页数:16
相关论文
共 45 条
[1]  
Chung J., Gulcehre C., Cho K., Bengio Y., Empirical evaluation of gated recurrent neural networks on sequence modeling, (2014)
[2]  
Kowsari K., Heidarysafa M., Brown D. E., Meimandi K. J., Barnes L. E., Rmdl: Random multimodel deep learning for classification, Proc. of the 2nd Int. Conf. on Information System and Data Mining, pp. 19-28, (2018)
[3]  
Robertson S., Understanding inverse document frequency: On theoretical arguments for IDF, Journal of Documentation, 60, 5, pp. 503-520, (2004)
[4]  
Wang B., Klabjan D., Regularization for unsupervised deep neural nets, Thirty-First AAAI Conf. on Artificial Intelligence, pp. 2681-2687, (2017)
[5]  
Kubo Y., Tucker G., Wiesler S., Compacting neural network classifiers via dropout training, (2016)
[6]  
Gal Y., Ghahramani Z., A theoretically grounded application of dropout in recurrent neural networks, Proc. of the Advances in Neural Information Processing Systems, pp. 1019-1027, (2016)
[7]  
Srivastava N., Hinton G., Krizhevsky A., Sutskever I., Salakhutdinov R., Dropout: A simple way to prevent neural networks from overfitting, The Journal of Machine Learning Research, 15, 1, pp. 1929-1958, (2014)
[8]  
Ioffe S., Szegedy C., Batch normalization: Accelerating deep network training by reducing internal covariate shift, Int. Conf. on Machine Learning, PMLR, Lillie, France, pp. 448-456, (2015)
[9]  
Xiang S., Li H., On the effects of batch and weight normalization in generative adversarial networks, (2017)
[10]  
Simard P.Y., Steinkras D., Platt J.C., Best practices for convolutional neural networks applied to visual document analysis, Proc. of the Seventh Int. Conf. on Document Analysis and Recognition, pp. 958-962, (2003)