Dual Extrapolation for Sparse Generalized Linear Models

Abstract : Generalized Linear Models (GLM) form a wide class of regression and classification models, where prediction is a function of a linear combination of the input variables. For statistical inference in high dimension, sparsity inducing regularizations have proven to be useful while offering statistical guarantees. However, solving the resulting optimization problems can be challenging: even for popular iterative algorithms such as coordinate descent, one needs to loop over a large number of variables. To mitigate this, techniques known as screening rules and working sets diminish the size of the optimization problem at hand, either by progressively removing variables, or by solving a growing sequence of smaller problems. For both techniques, significant variables are identified thanks to convex duality arguments. In this paper, we show that the dual iterates of a GLM exhibit a Vector AutoRegressive (VAR) behavior after sign identification, when the primal problem is solved with proximal gradient descent or cyclic coordinate descent. Exploiting this regularity, one can construct dual points that offer tighter certificates of optimality, enhancing the performance of screening rules and helping to design competitive working set algorithms.
Document type :
Preprints, Working Papers, ...
Complete list of metadatas

Cited literature [64 references]  Display  Hide  Download

https://hal.archives-ouvertes.fr/hal-02263500
Contributor : Mathurin Massias <>
Submitted on : Monday, August 5, 2019 - 8:21:13 AM
Last modification on : Wednesday, August 7, 2019 - 1:11:59 AM

File

main.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-02263500, version 1

Citation

Mathurin Massias, Samuel Vaiter, Alexandre Gramfort, Joseph Salmon. Dual Extrapolation for Sparse Generalized Linear Models. 2019. ⟨hal-02263500⟩

Share

Metrics

Record views

30

Files downloads

39