arXiv Analytics

Sign in

arXiv:1802.08009 [cs.LG]AbstractReferencesReviewsResources

Iterate averaging as regularization for stochastic gradient descent

Gergely Neu, Lorenzo Rosasco

Published 2018-02-22Version 1

We propose and analyze a variant of the classic Polyak-Ruppert averaging scheme, broadly used in stochastic gradient methods. Rather than a uniform average of the iterates, we consider a weighted average, with weights decaying in a geometric fashion. In the context of linear least squares regression, we show that this averaging scheme has a the same regularizing effect, and indeed is asymptotically equivalent, to ridge regression. In particular, we derive finite-sample bounds for the proposed approach that match the best known results for regularized stochastic gradient methods.

Related articles: Most relevant | Search more
arXiv:2307.03886 [cs.LG] (Published 2023-07-08)
On Regularization and Inference with Label Constraints
arXiv:1411.1134 [cs.LG] (Published 2014-11-05)
Global Convergence of Stochastic Gradient Descent for Some Nonconvex Matrix Problems
arXiv:1212.1824 [cs.LG] (Published 2012-12-08, updated 2012-12-28)
Stochastic Gradient Descent for Non-smooth Optimization: Convergence Results and Optimal Averaging Schemes