Harder, Better, Faster, Stronger Convergence Rates for Least-Squares Regression - Archive ouverte HAL Accéder directement au contenu
Pré-Publication, Document De Travail Année : 2016

Harder, Better, Faster, Stronger Convergence Rates for Least-Squares Regression

Résumé

We consider the optimization of a quadratic objective function whose gradients are only accessible through a stochastic oracle that returns the gradient at any given point plus a zero-mean finite variance random error. We present the first algorithm that achieves jointly the optimal prediction error rates for least-squares regression, both in terms of forgetting of initial conditions in O(1/n2), and in terms of dependence on the noise and dimension d of the problem, as O(d/n). Our new algorithm is based on averaged accelerated regularized gradient descent, and may also be analyzed through finer assumptions on initial conditions and the Hessian matrix, leading to dimension-free quantities that may still be small while the “optimal” terms above are large. In order to characterize the tightness of these new bounds, we consider an application to non- parametric regression and use the known lower bounds on the statistical performance (without computational limits), which happen to match our bounds obtained from a single pass on the data and thus show optimality of our algorithm in a wide variety of particular trade-offs between bias and variance.
Fichier principal
Vignette du fichier
regularizedstrongconvex - non uniform.pdf (231.39 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-01275431 , version 1 (17-02-2016)
hal-01275431 , version 2 (23-02-2016)

Identifiants

Citer

Aymeric Dieuleveut, Nicolas Flammarion, Francis Bach. Harder, Better, Faster, Stronger Convergence Rates for Least-Squares Regression. 2016. ⟨hal-01275431v1⟩
433 Consultations
1709 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More