To Combine or Not to Combine? The Influence of Combining Training Datasets on the Robustness of Deep Learning Models: An Analysis for Optical Character Recognition of Handwriting

被引:0
作者
Fischer-Brandies, Leopold [1 ]
Mueller, Lucas [1 ]
Rebholz, Benjamin [1 ]
Buettner, Ricardo [1 ]
机构
[1] Helmut Schmidt Univ, Univ Fed Armed Forces Hamburg, Chair Hybrid Intelligence, D-22043 Hamburg, Germany
关键词
Robustness; Text recognition; Training data; Deep learning; Optical character recognition; Training; Data models; Accuracy; Handwriting recognition; Convolutional neural networks; training data; deep learning; transfer learning; OCR; handwriting; NEURAL-NETWORK; ONLINE;
D O I
10.1109/ACCESS.2025.3556582
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The present manuscript addresses the question of how training data should be sampled for deep learning models by analyzing and evaluating the impact of training data representation and complexity on the performance and robustness of deep learning models. To address this open question, we take a combinatorial approach and train three architecturally identical deep learning models on three combinations of handwritten digit datasets of varying complexity: EMNIST Digits, DIDA, and a newly composed third dataset combining the first two datasets. Each model was evaluated using withheld test data from all three datasets. We find that models trained exclusively on either EMNIST Digits or DIDA performed well on their respective datasets but poorly on unfamiliar datasets. However, the model trained on both datasets showed an overall solid performance, although not quite reaching the accuracy of the specialized models on their respective datasets. We conclude that while specializing in the training dataset can increase accuracy, a more diverse dataset enhances model robustness. In practice, deep learning models should thus be trained on data that represents the actual application environment as closely as possible or, if such data is not available, on diverse data.
引用
收藏
页码:59039 / 59056
页数:18
相关论文
共 82 条
[1]  
Abadi Martin, 2015, TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems
[2]   Medical Handwritten Prescription Recognition Using CRNN [J].
Achkar, Roger ;
Ghayad, Khodor ;
Haidar, Rayan ;
Saleh, Sawsan ;
Al Hajj, Rana .
PROCEEDING OF THE 2019 INTERNATIONAL CONFERENCE ON COMPUTER, INFORMATION AND TELECOMMUNICATION SYSTEMS (IEEE CITS 2019), 2019, :211-215
[3]   Performance of deep learning models for classifying and detecting common weeds in corn and soybean production systems [J].
Ahmad, Aanis ;
Saraswat, Dharmendra ;
Aggarwal, Varun ;
Etienne, Aaron ;
Hancock, Benjamin .
COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2021, 184 (184)
[4]   Offline handwritten Arabic cursive text recognition using Hidden Markov Models and re-ranking [J].
AlKhateeb, Jawad H. ;
Ren, Jinchang ;
Jiang, Jianmin ;
Al-Muhtaseb, Husni .
PATTERN RECOGNITION LETTERS, 2011, 32 (08) :1081-1088
[5]   Review of deep learning: concepts, CNN architectures, challenges, applications, future directions [J].
Alzubaidi, Laith ;
Zhang, Jinglan ;
Humaidi, Amjad J. ;
Al-Dujaili, Ayad ;
Duan, Ye ;
Al-Shamma, Omran ;
Santamaria, J. ;
Fadhel, Mohammed A. ;
Al-Amidie, Muthana ;
Farhan, Laith .
JOURNAL OF BIG DATA, 2021, 8 (01)
[6]   On-line handwriting recognition with support vector machines - A kernel approach [J].
Bahlmann, C ;
Haasdonk, B ;
Burkhardt, H .
EIGHTH INTERNATIONAL WORKSHOP ON FRONTIERS IN HANDWRITING RECOGNITION: PROCEEDINGS, 2002, :49-54
[7]   A Survey of Handwritten Character Recognition with MNIST and EMNIST [J].
Baldominos, Alejandro ;
Saez, Yago ;
Isasi, Pedro .
APPLIED SCIENCES-BASEL, 2019, 9 (15)
[8]  
Bojarski M, 2016, Arxiv, DOI arXiv:1604.07316
[9]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[10]   Deep Learning-Based Crack Damage Detection Using Convolutional Neural Networks [J].
Cha, Young-Jin ;
Choi, Wooram ;
Buyukozturk, Oral .
COMPUTER-AIDED CIVIL AND INFRASTRUCTURE ENGINEERING, 2017, 32 (05) :361-378