{ "id": "2205.15549", "version": "v1", "published": "2022-05-31T05:50:02.000Z", "updated": "2022-05-31T05:50:02.000Z", "title": "VC Theoretical Explanation of Double Descent", "authors": [ "Eng Hock Lee", "Vladimir Cherkassky" ], "categories": [ "stat.ML", "cs.LG" ], "abstract": "There has been growing interest in generalization performance of large multilayer neural networks that can be trained to achieve zero training error, while generalizing well on test data. This regime is known as 'second descent' and it appears to contradict conventional view that optimal model complexity should reflect optimal balance between underfitting and overfitting, aka the bias-variance trade-off. This paper presents VC-theoretical analysis of double descent and shows that it can be fully explained by classical VC generalization bounds. We illustrate an application of analytic VC-bounds for modeling double descent for classification problems, using empirical results for several learning methods, such as SVM, Least Squares, and Multilayer Perceptron classifiers. In addition, we discuss several possible reasons for misinterpretation of VC-theoretical results in the machine learning community.", "revisions": [ { "version": "v1", "updated": "2022-05-31T05:50:02.000Z" } ], "analyses": { "keywords": [ "double descent", "vc theoretical explanation", "large multilayer neural networks", "achieve zero training error", "classical vc generalization bounds" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }