arXiv Analytics

Sign in

arXiv:1810.09418 [cs.LG]AbstractReferencesReviewsResources

Optimality of the final model found via Stochastic Gradient Descent

Andrea Schioppa

Published 2018-10-22Version 1

We study convergence properties of Stochastic Gradient Descent (SGD) for convex objectives without assumptions on smoothness or strict convexity. We consider the question of establishing that with high probability the objective evaluated at the candidate minimizer returned by SGD is close to the minimal value of the objective. We compare this result concerning the final candidate minimzer (i.e. the final model parameters learned after all gradient steps) to the online learning techniques of [Zin03] that take a rolling average of the model parameters at the different steps of SGD.

Related articles: Most relevant | Search more
arXiv:1411.1134 [cs.LG] (Published 2014-11-05)
Global Convergence of Stochastic Gradient Descent for Some Nonconvex Matrix Problems
arXiv:1509.09002 [cs.LG] (Published 2015-09-30)
Convergence of Stochastic Gradient Descent for PCA
arXiv:1212.1824 [cs.LG] (Published 2012-12-08, updated 2012-12-28)
Stochastic Gradient Descent for Non-smooth Optimization: Convergence Results and Optimal Averaging Schemes