Sparse recovery by reduced variance stochastic approximation - Archive ouverte HAL Accéder directement au contenu
Article Dans Une Revue Information and Inference Année : 2023

Sparse recovery by reduced variance stochastic approximation

Résumé

In this paper, we discuss application of iterative Stochastic Optimization routines to the problem of sparse signal recovery from noisy observation. Using Stochastic Mirror Descent algorithm as a building block, we develop a multistage procedure for recovery of sparse solutions to Stochastic Optimization problem under assumption of smoothness and quadratic minoration on the expected objective. An interesting feature of the proposed algorithm is its linear convergence of the approximate solution during the preliminary phase of the routine when the component of stochastic error in the gradient observation which is due to bad initial approximation of the optimal solution is larger than the "ideal" asymptotic error component owing to observation noise "at the optimal solution." We also show how one can straightforwardly enhance reliability of the corresponding solution by using Median-of-Means like techniques. We illustrate the performance of the proposed algorithms in application to classical problems of recovery of sparse and low rank signals in linear regression framework. We show, under rather weak assumption on the regressor and noise distributions, how they lead to parameter estimates which obey (up to factors which are logarithmic in problem dimension and confidence level) the best known to us accuracy bounds.

Dates et versions

hal-03185516 , version 1 (30-03-2021)

Identifiants

Citer

Anatoli B. Juditsky, Andrei Kulunchakov, Hlib Tsyntseus. Sparse recovery by reduced variance stochastic approximation. Information and Inference, 2023, 12 (2), pp.851-896. ⟨10.1093/imaiai/iaac028⟩. ⟨hal-03185516⟩
121 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More