S. Arlot and P. Massart, Data-driven calibration of penalties for least-squares regression , technical report, 2008.

M. S. Asif and J. Romberg, Dantzig selector homotopy with dynamic measurements, Proceedings of SPIE Computational Imaging VII, 2009.

P. Bickel, Y. Ritov, and A. Tsybakov, Simultaneous analysis of Lasso and Dantzig selector, The Annals of Statistics, vol.37, issue.4, 2007.
DOI : 10.1214/08-AOS620

URL : https://hal.archives-ouvertes.fr/hal-00401585

L. Birgé, Model selection for density estimation with L 2 -loss, 2008.

L. Birgé and P. Massart, Minimal Penalties for Gaussian Model Selection, Probability Theory and Related Fields, vol.6, issue.1-2, pp.33-73, 2007.
DOI : 10.1007/s00440-006-0011-8

F. Bunea, A. B. Tsybakov, and M. H. Wegkamp, Aggregation and Sparsity Via ???1 Penalized Least Squares, Proceedings of 19th Annual Conference on Learning Theory (COLT 2006), Lecture Notes in Artificial Intelligence v.4005 (, 2006.
DOI : 10.1007/11776420_29

URL : https://hal.archives-ouvertes.fr/hal-00084553

F. Bunea, A. B. Tsybakov, and M. H. Wegkamp, Sparse Density Estimation with ???1 Penalties, Lecture Notes in Artificial Intelligence, vol.4539, pp.530-543, 2007.
DOI : 10.1007/978-3-540-72927-3_38

URL : https://hal.archives-ouvertes.fr/hal-00160850

F. Bunea, A. B. Tsybakov, and M. H. Wegkamp, Aggregation for Gaussian regression, The Annals of Statistics, vol.35, issue.4, pp.1674-1697, 2007.
DOI : 10.1214/009053606000001587

F. Bunea, A. B. Tsybakov, and M. H. Wegkamp, Sparsity oracle inequalities for the Lasso, Electronic Journal of Statistics, vol.1, issue.0, pp.169-194, 2007.
DOI : 10.1214/07-EJS008

URL : https://hal.archives-ouvertes.fr/hal-00160646

F. Bunea, A. B. Tsybakov, and M. H. Wegkamp, SPADES and mixture models, The Annals of Statistics, vol.38, issue.4, 2009.
DOI : 10.1214/09-AOS790

URL : https://hal.archives-ouvertes.fr/hal-00514124

F. Bunea, Consistent selection via the Lasso for high dimensional approximating regression models, IMS Lecture notes-Monograph Series, pp.122-137, 2008.

E. J. Candès and . Y. Plan, Near-ideal model selection by l1 minimization, 2007.

E. J. Candès and T. Tao, The Dantzig selector: Statistical estimation when p is much larger than n, The Annals of Statistics, vol.35, issue.6, pp.2313-2351, 2007.
DOI : 10.1214/009053606000001523

D. Chen, D. L. Donoho, and M. Saunders, Atomic decomposition by basis pursuit, SIAM review, pp.129-159, 2001.

D. L. Donoho, M. Elad, and V. Temlyakov, Stable recovery of sparse overcomplete representations in the presence of noise, IEEE Transactions on Information Theory, vol.52, issue.1, pp.6-18, 2006.
DOI : 10.1109/TIT.2005.860430

D. L. Donoho and I. M. Johnstone, Ideal spatial adaptation by wavelet shrinkage, Biometrika, vol.81, issue.3, pp.425-455, 1994.
DOI : 10.1093/biomet/81.3.425

A. Juditsky and S. Lambert-lacroix, On minimax density estimation on \mathbb{R}}, Bernoulli, vol.10, issue.2, pp.187-220, 2004.
DOI : 10.3150/bj/1082380217

K. Knight and W. Fu, Asymptotics for lasso-type estimators, Ann. Statist, vol.28, issue.5, pp.1356-1378, 2000.

K. Lounici, Sup-norm convergence rate and sign concentration property of Lasso and Dantzig estimators, Electronic Journal of Statistics, vol.2, issue.0, 2008.
DOI : 10.1214/08-EJS177

URL : https://hal.archives-ouvertes.fr/hal-00222251

P. Massart, Concentration inequalities and model selection. Lectures from the 33rd Summer School on Probability Theory held in Saint-Flour, 2003.

N. Meinshausen and P. Buhlmann, High-dimensional graphs and variable selection with the Lasso, The Annals of Statistics, vol.34, issue.3, pp.1436-1462, 2006.
DOI : 10.1214/009053606000000281

N. Meinhausen and B. Yu, Lasso-type recovery of sparse representations for high-dimensional data, The Annals of Statistics, vol.37, issue.1, pp.246-270, 2009.
DOI : 10.1214/07-AOS582

M. R. Osborne, B. Presnell, and B. A. Turlach, On the Lasso and its dual, Journal of Computational and Graphical Statistics, vol.9, pp.319-337, 2000.

M. R. Osborne, B. Presnell, and B. A. Turlach, A new approach to variable selection in least squares problems, IMA Journal of Numerical Analysis, vol.20, issue.3, pp.389-404, 2000.
DOI : 10.1093/imanum/20.3.389

P. Reynaud-bouret and V. Rivoirard, Calibration of thresholding rules for Poisson intensity estimation, p.1148, 2009.

P. Reynaud-bouret, V. Rivoirard, and C. Tuleau, On the influence of the support of functions for density estimation, 2009.

R. Tibshirani, Regression shrinkage and selection via the Lasso, Journal of the Royal Statistics Society, Series B, vol.58, pp.267-288, 1996.

S. Van-de-geer, High-dimensional generalized linear models and the lasso, The Annals of Statistics, vol.36, issue.2, pp.614-645, 2008.
DOI : 10.1214/009053607000000929

B. Yu and P. Zhao, On model selection consistency of Lasso estimators, Journal of Machine Learning Research, vol.7, pp.2541-2567, 2006.

C. H. Zhang and J. Huang, The sparsity and bias of the Lasso selection in highdimensional linear regression, 2007.

H. Zou, The Adaptive Lasso and Its Oracle Properties, Journal of the American Statistical Association, vol.101, issue.476, pp.1418-1429, 2006.
DOI : 10.1198/016214506000000735