D. Achlioptas, Database-friendly random projections: Johnson-Lindenstrauss with binary coins, Journal of Computer and System Sciences, vol.66, issue.4, pp.671-687, 2003.
DOI : 10.1016/S0022-0000(03)00025-4

N. Ailon and B. Chazelle, Approximate nearest neighbors and the fast Johnson- Lindenstrauss transform, STOC '06: Proceedings of the thirty-eighth annual ACM symposium on Theory of computing, pp.557-563, 2006.

J. Audibert and O. Catoni, Risk bounds in linear regression through pac-bayesian truncation, 2009.
URL : https://hal.archives-ouvertes.fr/hal-00360268

D. Bau, I. , and L. N. Trefethen, Numerical linear algebra, 1997.

J. Peter, . Bickel, A. B. Ritov, and . Tsybakov, Simultaneous analysis of Lasso and Dantzig selector, 2008.

A. Blum, Random projection, margins, kernels, and feature-selection. Subspace, Latent Structure and Feature Selection, pp.52-68, 2006.

R. Calderbank, S. Jafarpour, and R. Schapire, Compressed learning: Universal sparse dimensionality reduction and learning in the measurement domain, 2009.

E. Candes and T. Tao, The Dantzig selector: Statistical estimation when p is much larger than n, The Annals of Statistics, vol.35, issue.6, p.2313, 2007.
DOI : 10.1214/009053606000001523

J. Emmanuel, J. K. Candes, and . Romberg, Signal recovery from random projections, SPIE, vol.5674, pp.76-86, 2005.

S. S. Chen, D. L. Donoho, and M. A. Saunders, Atomic Decomposition by Basis Pursuit, SIAM Journal on Scientific Computing, vol.20, issue.1, pp.33-61, 1998.
DOI : 10.1137/S1064827596304010

A. Mark, M. B. Davenport, R. G. Wakin, and . Baraniuk, Detection and estimation with compressive measurements, 2006.

E. Greenshtein and Y. Ritov, Persistence in high-dimensional linear predictor selection and the virtue of overparametrization, Bernoulli, vol.10, issue.6, pp.971-988, 2004.
DOI : 10.3150/bj/1106314846

L. Györfi, M. Kohler, A. Krzy?, and H. Walk, A distribution-free theory of nonparametric regression, 2002.
DOI : 10.1007/b97848

M. Sham, K. Kakade, A. Sridharan, and . Tewari, On the complexity of linear prediction: Risk bounds, margin bounds, and regularization, Neural Information Processing Systems, pp.793-800, 2008.

S. Mallat, A Wavelet Tour of Signal Processing, 1999.

Y. Nardi and A. Rinaldo, On the asymptotic properties of the group lasso estimator for linear models, Electronic Journal of Statistics, vol.2, issue.0, pp.605-633, 2008.
DOI : 10.1214/08-EJS200

D. Pollard, Convergence of Stochastic Processes, 1984.
DOI : 10.1007/978-1-4612-5254-2

A. Rahimi and B. Recht, Random features for large-scale kernel machines, Neural Information Processing Systems, 2007.

S. Rosset and J. Zhu, Piecewise linear regularized solution paths, The Annals of Statistics, vol.35, issue.3, p.1012, 2007.
DOI : 10.1214/009053606000001370

URL : http://arxiv.org/abs/0708.2197

R. Tibshirani, Regression shrinkage and selection via the Lasso, Journal of the Royal Statistical Society, Series B, vol.58, pp.267-288, 1994.

A. N. Tikhonov, Solution of incorrectly formulated problems and the regularization method, Soviet Math Dokl, vol.4, pp.1035-1038, 1963.

Y. Tsaig and D. L. Donoho, Compressed sensing, IEEE Trans. Inform. Theory, vol.52, pp.1289-1306, 2006.

N. Vladimir and . Vapnik, The nature of statistical learning theory, 1995.

T. Zhang, Covering number bounds of certain regularized linear function classes, Journal of Machine Learning Research, vol.2, pp.527-550, 2002.

T. Zhang, Some sharp performance bounds for least squares regression with L1 regularization, 2009.

S. Zhou, J. D. Lafferty, and L. A. Wasserman, Compressed regression, Neural Information Processing Systems, 2007.