{ "id": "2409.07434", "version": "v1", "published": "2024-09-11T17:28:38.000Z", "updated": "2024-09-11T17:28:38.000Z", "title": "Asymptotics of Stochastic Gradient Descent with Dropout Regularization in Linear Models", "authors": [ "Jiaqi Li", "Johannes Schmidt-Hieber", "Wei Biao Wu" ], "comment": "77 pages, 5 figures, 4 tables", "categories": [ "stat.ML", "cs.LG", "math.ST", "stat.TH" ], "abstract": "This paper proposes an asymptotic theory for online inference of the stochastic gradient descent (SGD) iterates with dropout regularization in linear regression. Specifically, we establish the geometric-moment contraction (GMC) for constant step-size SGD dropout iterates to show the existence of a unique stationary distribution of the dropout recursive function. By the GMC property, we provide quenched central limit theorems (CLT) for the difference between dropout and $\\ell^2$-regularized iterates, regardless of initialization. The CLT for the difference between the Ruppert-Polyak averaged SGD (ASGD) with dropout and $\\ell^2$-regularized iterates is also presented. Based on these asymptotic normality results, we further introduce an online estimator for the long-run covariance matrix of ASGD dropout to facilitate inference in a recursive manner with efficiency in computational time and memory. The numerical experiments demonstrate that for sufficiently large samples, the proposed confidence intervals for ASGD with dropout nearly achieve the nominal coverage probability.", "revisions": [ { "version": "v1", "updated": "2024-09-11T17:28:38.000Z" } ], "analyses": { "subjects": [ "62E20", "62F12", "68W27" ], "keywords": [ "stochastic gradient descent", "dropout regularization", "linear models", "constant step-size sgd dropout iterates", "quenched central limit theorems" ], "note": { "typesetting": "TeX", "pages": 77, "language": "en", "license": "arXiv", "status": "editable" } } }