arXiv Analytics

Sign in

arXiv:2311.14079 [cs.LG]AbstractReferencesReviewsResources

Empirical Comparison between Cross-Validation and Mutation-Validation in Model Selection

Jinyang Yu, Sami Hamdan, Leonard Sasse, Abigail Morrison, Kaustubh R. Patil

Published 2023-11-23Version 1

Mutation validation (MV) is a recently proposed approach for model selection, garnering significant interest due to its unique characteristics and potential benefits compared to the widely used cross-validation (CV) method. In this study, we empirically compared MV and $k$-fold CV using benchmark and real-world datasets. By employing Bayesian tests, we compared generalization estimates yielding three posterior probabilities: practical equivalence, CV superiority, and MV superiority. We also evaluated the differences in the capacity of the selected models and computational efficiency. We found that both MV and CV select models with practically equivalent generalization performance across various machine learning algorithms and the majority of benchmark datasets. MV exhibited advantages in terms of selecting simpler models and lower computational costs. However, in some cases MV selected overly simplistic models leading to underfitting and showed instability in hyperparameter selection. These limitations of MV became more evident in the evaluation of a real-world neuroscientific task of predicting sex at birth using brain functional connectivity.

Related articles: Most relevant | Search more
arXiv:2111.04688 [cs.LG] (Published 2021-11-08, updated 2022-06-30)
Universal and data-adaptive algorithms for model selection in linear contextual bandits
arXiv:1909.07140 [cs.LG] (Published 2019-09-16)
Weighted Sampling for Combined Model Selection and Hyperparameter Tuning
arXiv:2409.09674 [cs.LG] (Published 2024-09-15)
Model Selection Through Model Sorting