Model validation using mutated training labels: An exploratory study

被引:4
作者
Zhang, Jie M. [1 ]
Harman, Mark [2 ,7 ]
Guedj, Benjamin [3 ,4 ,8 ]
Barr, Earl T. [5 ]
Shawe-Taylor, John [6 ]
机构
[1] Kings Coll London, Engn, Dept Informat, London, England
[2] UCL, CREST Ctr, London, England
[3] UCL, Ctr Artificial Intelligence, London, England
[4] UCL, Dept Comp Sci, London, England
[5] UCL, Software Engn, London, England
[6] UCL, London, England
[7] Meta Platforms, Software Engn Automat, London, England
[8] Inria, Palaiseau, France
关键词
Model validation; Model complexity; Model overfitting; NOISE;
D O I
10.1016/j.neucom.2023.02.042
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We introduce an exploratory study on Mutation Validation (MV), a model validation method using mutated training labels for supervised learning. MV mutates training data labels, retrains the model against the mutated data, and then uses the metamorphic relation that captures the consequent training performance changes to assess model fit. It does not use a validation set or test set. The intuition under-pinning MV is that overfitting models tend to fit noise in the training data.MV does not aim to replace out-of-sample validation. Instead, we provide the first exploratory study on the possibility of using MV as a complement of out-of-sample validation. We explore 8 different learning algorithms, 18 datasets, and 5 types of hyperparameter tuning tasks. Our results demonstrate that MV complements well cross-validation and test accuracy in model selection and hyperparameter tuning tasks. MV deserves more attention from developers when simplicity, sustainaiblity, security (e.g., defend-ing training data attack), and interpretability of the built models are required.(c) 2023 Published by Elsevier B.V.
引用
收藏
页数:12
相关论文
共 21 条
[1]  
[Anonymous], 2019, CLASSIFIER COM
[2]  
Asuncion A., 2007, UCI MACHINE LEARNING
[3]  
Chen TY, 2020, Arxiv, DOI arXiv:2002.12543
[4]  
Feldman V, 2019, PR MACH LEARN RES, V97
[5]  
Ghosh A, 2017, AAAI CONF ARTIF INTE, P1919
[6]   LSTM: A Search Space Odyssey [J].
Greff, Klaus ;
Srivastava, Rupesh K. ;
Koutnik, Jan ;
Steunebrink, Bas R. ;
Schmidhuber, Juergen .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (10) :2222-2232
[7]   Limitations of Bayesian Leave-One-Out Cross-Validation for Model Selection [J].
Gronau Q.F. ;
Wagenmakers E.-J. .
Computational Brain & Behavior, 2019, 2 (1) :1-11
[8]   The problem of overfitting [J].
Hawkins, DM .
JOURNAL OF CHEMICAL INFORMATION AND COMPUTER SCIENCES, 2004, 44 (01) :1-12
[9]   USING ADDITIVE NOISE IN BACK-PROPAGATION TRAINING [J].
HOLMSTROM, L ;
KOISTINEN, P .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1992, 3 (01) :24-38
[10]  
Goodfellow IJ, 2015, Arxiv, DOI arXiv:1412.6572