arXiv Analytics

Sign in

arXiv:2006.07002 [cs.LG]AbstractReferencesReviewsResources

Double Double Descent: On Generalization Errors in Transfer Learning between Linear Regression Tasks

Yehuda Dar, Richard G. Baraniuk

Published 2020-06-12Version 1

We study the transfer learning process between two linear regression problems. An important and timely special case is when the regressors are overparameterized and perfectly interpolate their training data. We examine a parameter transfer mechanism whereby a subset of the parameters of the target task solution are constrained to the values learned for a related source task. We analytically characterize the generalization error of the target task in terms of the salient factors in the transfer learning architecture, i.e., the number of examples available, the number of (free) parameters in each of the tasks, the number of parameters transferred from the source to target task, and the correlation between the two tasks. Our non-asymptotic analysis shows that the generalization error of the target task follows a two-dimensional double descent trend (with respect to the number of free parameters in each of the tasks) that is controlled by the transfer learning factors. Our analysis points to specific cases where the transfer of parameters is beneficial.

Related articles: Most relevant | Search more
arXiv:2305.17297 [cs.LG] (Published 2023-05-26)
Generalization Error without Independence: Denoising, Linear Regression, and Transfer Learning
arXiv:1206.3274 [cs.LG] (Published 2012-06-13)
Small Sample Inference for Generalization Error in Classification Using the CUD Bound
arXiv:1711.05482 [cs.LG] (Published 2017-11-15)
Efficient Estimation of Generalization Error and Bias-Variance Components of Ensembles