R. J. Adler, An introduction to continuity, extrema, and related topics for general Gaussian processes, Institute of Mathematical Statistics Lecture Notes?Monograph Series, vol.12, 1990.

H. Akaike, Information theory and an extension of the maximum likelihood principle, Second International Symposium on Information Theory (Tsahkadsor, 1971), pp.267-281, 1973.

D. M. Allen, The Relationship Between Variable Selection and Data Agumentation and a Method for Prediction, Technometrics, vol.23, issue.3, pp.125-127, 1974.
DOI : 10.2307/1267352

P. Alquier, LASSO, Iterative Feature Selection and the Correlation Selector: Oracle inequalities and numerical performances, Electronic Journal of Statistics, vol.2, issue.0, pp.1129-1152, 2008.
DOI : 10.1214/08-EJS288

URL : https://hal.archives-ouvertes.fr/hal-00181784

P. Alquier, PAC-Bayesian bounds for randomized empirical risk minimizers, Mathematical Methods of Statistics, vol.17, issue.4, pp.279-304, 2008.
DOI : 10.3103/S1066530708040017

URL : https://hal.archives-ouvertes.fr/hal-00354922

P. Alquier, Iterative feature selection in least square regression estimation, Annales de l'Institut Henri Poincare (B) Probability and Statistics, vol.44, issue.1
DOI : 10.1214/07-AIHP106

URL : https://hal.archives-ouvertes.fr/hal-00013780

P. Alquier and M. Et-hebiri, Generalization of constraints for high dimensional regression problems, Statistics & Probability Letters, vol.81, issue.12, 2008.
DOI : 10.1016/j.spl.2011.07.011

URL : https://hal.archives-ouvertes.fr/hal-00336101

P. Alquier and M. Et-hebiri, Transductive extensions of the LASSO and the Dantzig Selector, 2009.

J. Y. Audibert, Fast learning rates in statistical inference through aggregation, The Annals of Statistics, vol.37, issue.4, 2008.
DOI : 10.1214/08-AOS623

URL : https://hal.archives-ouvertes.fr/hal-00139030

F. Bach, Bolasso, Proceedings of the 25th international conference on Machine learning, ICML '08, 2008.
DOI : 10.1145/1390156.1390161

URL : https://hal.archives-ouvertes.fr/hal-00271289

F. Bach, Model-consistent sparse estimation through the bootstrap. Manuscript [12] Bach F. Consistency of the group lasso and multiple kernel learning, J. Mach. Learn. Res, vol.9, pp.1179-1225, 2008.
URL : https://hal.archives-ouvertes.fr/hal-00354771

Y. Baraud, Model selection for regression on a random design, ESAIM: Probability and Statistics, vol.6, pp.127-146, 2002.
DOI : 10.1051/ps:2002007

A. Barron, L. Birgé, and P. Et-massart, Risk bounds for model selection via penalization. Probab. Theory Related Fields, pp.301-413, 1999.

P. Bauer, B. M. Pötscher, and P. Et-hackl, Model selection by multiple test procedures, Statistics, vol.6, issue.1, pp.39-44, 1988.
DOI : 10.1080/02331888808802068

Y. Benjamini and Y. Et-hochberg, Controlling the false discovery rate : a practical and powerful approach to multiple testing, J. Roy. Statist. Soc. Ser. B, vol.57, issue.1, pp.289-300, 1995.

P. Bickel, Y. Ritov, and A. Et-tsybakov, Simultaneous analysis of Lasso and Dantzig selector, The Annals of Statistics, vol.37, issue.4, 2007.
DOI : 10.1214/08-AOS620

URL : https://hal.archives-ouvertes.fr/hal-00401585

P. J. Bickel and B. Li, Regularization in statistics, Test, vol.67, issue.2, pp.271-344, 2006.
DOI : 10.1007/BF02607055

L. Birgé, Model selection via testing: an alternative to (penalized) maximum likelihood estimators, Annales de l'Institut Henri Poincare (B) Probability and Statistics, vol.42, issue.3, pp.273-325, 2006.
DOI : 10.1016/j.anihpb.2005.04.004

L. Birgé and P. Et-massart, Minimal penalties for Gaussian model selection. Probab. Theory Related Fields, pp.33-73, 2007.

L. Birgé and P. Et-massart, Gaussian model selection, Journal of the European Mathematical Society, vol.3, issue.3, pp.203-268, 2001.
DOI : 10.1007/s100970100031

L. Breiman, Heuristics of instability and stabilization in model selection, The Annals of Statistics, vol.24, issue.6, pp.2350-2383, 1996.
DOI : 10.1214/aos/1032181158

L. Breiman, Better Subset Regression Using the Nonnegative Garrote, Technometrics, vol.37, issue.4, pp.373-384, 1995.
DOI : 10.1080/01621459.1980.10477428

P. Bühlmann and T. Et-hothorn, Twin Boosting: improved feature selection and prediction, Statistics and Computing, vol.67, issue.2, 2009.
DOI : 10.1007/s11222-009-9148-5

P. Bühlmann and B. Et-yu, Invited discussion on three papers on boosting by j.g, lugosi and vayatis, and zhang, 2004.

F. Bunea, Honest variable selection in linear and logistic regression models via ???1 and ???1+???2 penalization, Electronic Journal of Statistics, vol.2, issue.0, pp.1153-1194, 2008.
DOI : 10.1214/08-EJS287

F. Bunea, Consistent covariate selection and post model selection inference in semiparametric regression, The Annals of Statistics, vol.32, issue.3, pp.898-927, 2004.
DOI : 10.1214/009053604000000247

F. Bunea, Consistent selection via the Lasso for high dimensional approximating regression models, IMS Collections, 2008.
DOI : 10.1214/074921708000000101

F. Bunea, A. Tsybakov, and M. H. Et-wegkamp, Aggregation for regression learning. manuscript, 2004.
URL : https://hal.archives-ouvertes.fr/hal-00003205

F. Bunea, A. B. Tsybakov, and M. H. Et-wegkamp, Aggregation and Sparsity Via ???1 Penalized Least Squares, Learning theory
DOI : 10.1007/11776420_29

URL : https://hal.archives-ouvertes.fr/hal-00084553

F. Bunea, M. H. Wegkamp, and A. Et-auguste, Consistent variable selection in high dimensional regression via multiple testing, Journal of Statistical Planning and Inference, vol.136, issue.12, pp.4349-4364, 2006.
DOI : 10.1016/j.jspi.2005.03.011

F. Bunea, A. Tsybakov, and M. Et-wegkamp, Aggregation for Gaussian regression, The Annals of Statistics, vol.35, issue.4, pp.1674-1697, 2007.
DOI : 10.1214/009053606000001587

F. Bunea, A. B. Tsybakov, and M. H. Et-wegkamp, Sparsity oracle inequalities for the Lasso, Electronic Journal of Statistics, vol.1, issue.0, pp.169-194, 2007.
DOI : 10.1214/07-EJS008

URL : https://hal.archives-ouvertes.fr/hal-00160646

T. Cai, On block thresholding in wavelet regression : adaptivity, block size and threshold level, Statist. Sinica, vol.12, issue.4, pp.1241-1273, 2002.

E. Candès, T. Et-tao, and . Rejoinder, The Dantzig selector: Statistical estimation when p is much larger than n, The Annals of Statistics, vol.35, issue.6, pp.2313-23512392, 2007.
DOI : 10.1214/009053606000001523

G. Casella and R. L. Et-berger, Statistical inference. The Wadsworth & Brooks/Cole Statistics/Probability Series, 1990.

O. Catoni, A pac-bayesian approach to adaptive classification, 2003.

O. Catoni, Statistical learning theory and stochastic optimization, volume 1851 of Lecture Notes in Mathematics, Lecture notes from the 31st Summer School on Probability Theory held in Saint-Flour, 2001.

O. Catoni, Pac-bayesian supervised classification : The thermodynamics of statistical learning, IMS Lecture Notes Monograph Series, vol.56, 2007.
URL : https://hal.archives-ouvertes.fr/hal-00206119

L. Cavalier and A. B. Et-tsybakov, Penalized blockwise Stein's method, monotone oracles and sharp adaptive estimation, Math. Methods Statist. Meeting on Mathematical Statistics, vol.10, issue.3, pp.247-282, 2000.

O. Chapelle, B. Schölkopf, and A. Et-zien, Semi-supervised learning, 2006.
DOI : 10.7551/mitpress/9780262033589.001.0001

S. S. Chen and D. L. Et-donoho, Atomic decomposition by basis pursuit, 1995.

S. S. Chen, D. L. Donoho, and S. M. , Atomic decomposition by basis pursuit

C. Chesneau and M. Et-hebiri, Some theoretical results on the Grouped Variables Lasso, Mathematical Methods of Statistics, vol.17, issue.4, pp.317-326, 2008.
DOI : 10.3103/S1066530708040030

URL : https://hal.archives-ouvertes.fr/hal-00145160

P. G. Ciarlet, Introduction à l'analyse numérique matricielle et à l'optimisation, Collection Mathématiques Appliquées pour la Maîtrise. [Collection of Applied Mathematics for the Master's Degree]. Masson, 1982.

A. S. Dalalyan and A. B. Et-tsybakov, Aggregation by Exponential Weighting and Sharp Oracle Inequalities, Learning theory, pp.97-111, 2007.
DOI : 10.1007/978-3-540-72927-3_9

URL : https://hal.archives-ouvertes.fr/hal-00160857

A. S. Dalalyan and A. B. Et-tsybakov, Aggregation by exponential weighting, sharp PAC-Bayesian bounds and sparsity, Machine Learning, vol.52, issue.1-2, pp.39-61, 2008.
DOI : 10.1007/s10994-008-5051-0

URL : https://hal.archives-ouvertes.fr/hal-00265651

A. S. Dalalyan and A. B. Et-tsybakov, Sparse regression learning by aggregation and langevin monte-carlo. Manuscript, 2009.
DOI : 10.1016/j.jcss.2011.12.023

URL : https://hal.archives-ouvertes.fr/hal-00362471

I. Daubechies, M. Defrise, and C. Et-de-mol, An iterative thresholding algorithm for linear inverse problems with a sparsity constraint, Communications on Pure and Applied Mathematics, vol.58, issue.11, pp.1413-1457, 2004.
DOI : 10.1002/cpa.20042

D. Mol, C. , D. Vito, E. Et-rosasco, and L. , Elastic-net regularization in learning theory, 2008.

D. Donoho, M. Elad, and V. Et-temlyakov, Stable recovery of sparse overcomplete representations in the presence of noise, IEEE Transactions on Information Theory, vol.52, issue.1, pp.6-18, 2006.
DOI : 10.1109/TIT.2005.860430

D. L. Donoho, For most large underdetermined systems of linear equations the minimal ???1-norm solution is also the sparsest solution, Communications on Pure and Applied Mathematics, vol.50, issue.6, pp.797-829, 2006.
DOI : 10.1002/cpa.20132

D. L. Donoho and I. M. Et-johnstone, Adapting to Unknown Smoothness via Wavelet Shrinkage, Journal of the American Statistical Association, vol.31, issue.432, pp.1200-1224, 1995.
DOI : 10.1080/01621459.1979.10481038

D. L. Donoho and I. M. Et-johnstone, Ideal spatial adaptation by wavelet shrinkage, Biometrika, vol.81, issue.3, pp.425-455, 1994.
DOI : 10.1093/biomet/81.3.425

D. L. Donoho and J. Tanner, Sparse nonnegative solution of underdetermined linear equations by linear programming, Proc. Natl. Acad. Sci. USA, pp.9446-9451, 2005.
DOI : 10.1073/pnas.0502269102

D. L. Donoho, I. M. Johnstone, G. Kerkyacharian, and D. Et-picard, Wavelet shrinkage : asymptopia ?, J. Roy. Statist. Soc. Ser. B, vol.57, issue.2, pp.301-369, 1995.

B. Efron, Bootstrap Methods: Another Look at the Jackknife, The Annals of Statistics, vol.7, issue.1, pp.1-26, 1979.
DOI : 10.1214/aos/1176344552

B. Efron, How Biased is the Apparent Error Rate of a Prediction Rule?, Journal of the American Statistical Association, vol.39, issue.394, pp.461-470, 1986.
DOI : 10.1080/01621459.1986.10478291

B. Efron, The jackknife, the bootstrap and other resampling plans, CBMS- NSF Regional Conference Series in Applied Mathematics. Society for Industrial and Applied Mathematics (SIAM), vol.38, 1982.
DOI : 10.1137/1.9781611970319

B. Efron and R. J. Et-tibshirani, An introduction to the bootstrap, of Monographs on Statistics and Applied Probability. Chapman and Hall, 1993.
DOI : 10.1007/978-1-4899-4541-9

B. Efron, T. Hastie, I. Johnstone, and R. Et-tibshirani, Least angle regression, With discussion, and a rejoinder by the authors, pp.407-499, 2004.

J. Fan, Comments on ??Wavelets in statistics: A review?? by A. Antoniadis, Journal of the Italian Statistical Society, vol.58, issue.2, pp.131-138, 1997.
DOI : 10.1007/BF03178906

J. Fan and R. Li, Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties, Journal of the American Statistical Association, vol.96, issue.456, pp.1348-1360, 2001.
DOI : 10.1198/016214501753382273

J. Fan and H. Et-peng, Nonconcave penalized likelihood with a diverging number of parameters, Ann. Statist, vol.32, issue.3, pp.928-961, 2004.

D. P. Foster and E. I. Et-george, The Risk Inflation Criterion for Multiple Regression, The Annals of Statistics, vol.22, issue.4, pp.1947-1975, 1994.
DOI : 10.1214/aos/1176325766

I. E. Frank and J. H. Et-friedman, A Statistical View of Some Chemometrics Regression Tools, Technometrics, vol.5, issue.2, pp.109-135, 1993.
DOI : 10.1080/00401706.1993.10485033

J. H. Friedman, machine., The Annals of Statistics, vol.29, issue.5, pp.1189-1232, 2001.
DOI : 10.1214/aos/1013203451

W. J. Fu, Penalized Regressions: The Bridge versus the Lasso, Journal of Computational and Graphical Statistics, vol.58, issue.3, pp.397-416, 1998.
DOI : 10.1080/00401706.1993.10485033

P. Garrigues and L. Et-el-ghaoui, An homotopy algorithm for the lasso with online observations, To appear in Neural Information Processing Systems (NIPS) 21, 2008.

E. Greenshtein and Y. Et-ritov, Persistence in high-dimensional linear predictor selection and the virtue of overparametrization, Bernoulli, vol.10, issue.6, pp.971-988, 2004.
DOI : 10.3150/bj/1106314846

X. Guyon and J. Et-yao, On the Underfitting and Overfitting Sets of Models Chosen by Order Selection Criteria, Journal of Multivariate Analysis, vol.70, issue.2, pp.221-249, 1999.
DOI : 10.1006/jmva.1999.1828

URL : https://hal.archives-ouvertes.fr/hal-00272372

L. Györfi, M. Kohler, A. Krzy?ak, and H. Et-walk, A distribution-free theory of nonparametric regression, 2002.
DOI : 10.1007/b97848

T. Hastie, R. Tibshirani, and J. Et-friedman, The elements of statistical learning, Data mining, inference, and prediction, 2001.

T. J. Hastie and R. J. Et-tibshirani, Generalized additive models, volume 43 of Monographs on Statistics and Applied Probability, 1990.

D. M. Haughton, On the Choice of a Model to Fit Data from an Exponential Family, The Annals of Statistics, vol.16, issue.1
DOI : 10.1214/aos/1176350709

M. Hebiri, Sparse conformal predictors, Statistics and Computing, vol.35, issue.2, 2008.
DOI : 10.1007/s11222-009-9167-2

URL : https://hal.archives-ouvertes.fr/hal-00360771

M. Hebiri, Regularization with the smooth-lasso procedure, Preprint Laboratoire de Probabilités et Modèles Aléatoires, 2008.
URL : https://hal.archives-ouvertes.fr/hal-00260816

H. Arthur, E. Et-kennard, and R. W. , A note on a power generalization of ridge regression, Technometrics, vol.17, p.269, 1975.

C. Huang, G. L. Cheang, and A. Et-barron, Risk of penalized least squares, greedy selection and l1 penalization for flexible function libraries. preprint, 2008.

J. Huang and T. Et-zhang, The benefit of group sparsity. manuscript, 2009.

J. Huang, J. L. Horowitz, and S. Et-ma, Asymptotic properties of bridge estimators in sparse high-dimensional regression models, The Annals of Statistics, vol.36, issue.2, pp.587-613, 2008.
DOI : 10.1214/009053607000000875

D. R. Hunter and R. Li, Variable selection using MM algorithms, The Annals of Statistics, vol.33, issue.4, pp.1617-1642, 2005.
DOI : 10.1214/009053605000000200

G. M. James, P. Radchenko, and J. Et-lv, DASSO: connections between the Dantzig selector and lasso, Journal of the Royal Statistical Society: Series B (Statistical Methodology), vol.67, issue.1, pp.127-142, 2009.
DOI : 10.1111/j.1467-9868.2008.00668.x

J. Jia and B. Et-yu, On model selection consistency of the elastic net when p » n, 2008.

A. Juditsky and A. Et-nemirovski, Functional aggregation for nonparametric regression

A. Juditsky, P. Rigollet, and A. B. Et-tsybakov, Learning by mirror averaging, The Annals of Statistics, vol.36, issue.5, pp.2183-2206, 2008.
DOI : 10.1214/07-AOS546

URL : https://hal.archives-ouvertes.fr/hal-00341026

S. J. Kim, K. Koh, M. Lustig, S. Boyd, and D. Et-gorinevsky, An Interior-Point Method for Large-Scale -Regularized Least Squares, IEEE Journal of Selected Topics in Signal Processing, vol.1, issue.4, pp.606-617, 2007.
DOI : 10.1109/JSTSP.2007.910971

Y. Kim, J. Kim, and Y. Et-kim, Blockwise sparse regression, Statist. Sinica, vol.16, issue.2, pp.375-390, 2006.

K. Knight and W. Et-fu, Asymptotics for lasso-type estimators, Ann. Statist, vol.28, issue.5, pp.1356-1378, 2000.

V. Koltchinskii, Dantzig selector and sparsity oracle inequalities. Bernoulli, to appear, 2008.

V. Koltchinskii, Sparse recovery in convex hulls via entropy penalization, The Annals of Statistics, vol.37, issue.3, pp.1332-1359, 2009.
DOI : 10.1214/08-AOS621

V. Koltchinskii, Sparsity in penalized empirical risk minimization, Annales de l'Institut Henri Poincar??, Probabilit??s et Statistiques, vol.45, issue.1, 2007.
DOI : 10.1214/07-AIHP146

S. R. Land and J. H. Et-friedman, Variable fusion : a new method of adaptive signal regression, 1996.

S. L. Lauritzen, Graphical models, volume 17 of Oxford Statistical Science Series, 1996.

C. L. Lawson and R. J. Hanson, Solving least squares problems, Classics in Applied Mathematics. Society for Industrial and Applied Mathematics (SIAM), vol.15, 1995.
DOI : 10.1137/1.9781611971217

H. Leeb and B. M. Et-pötscher, MODEL SELECTION AND INFERENCE: FACTS AND FICTION, Econometric Theory, vol.307, issue.01, pp.21-59, 2005.
DOI : 10.1017/S0266466603191050__S0266466603191050

C. Leng, Y. Lin, and G. Et-wahba, A note on the lasso and related procedures in model selection, Statist. Sinica, vol.16, issue.4, pp.1273-1284, 2006.

G. Leung and A. R. Et-barron, Information Theory and Mixing Least-Squares Regressions, IEEE Transactions on Information Theory, vol.52, issue.8, pp.3396-3410, 2006.
DOI : 10.1109/TIT.2006.878172

K. Li and C. L. Asymptotic-optimality-for-c-p, Asymptotic Optimality for $C_p, C_L$, Cross-Validation and Generalized Cross-Validation: Discrete Index Set, The Annals of Statistics, vol.15, issue.3, pp.958-975, 1987.
DOI : 10.1214/aos/1176350486

K. Lounici, Sup-norm convergence rate and sign concentration property of Lasso and Dantzig estimators, Electronic Journal of Statistics, vol.2, issue.0, pp.90-102, 2008.
DOI : 10.1214/08-EJS177

URL : https://hal.archives-ouvertes.fr/hal-00222251

K. Lounici, M. Pontil, A. B. Tsybakov, and S. Et-van-de-geer, Taking advantage of sparsity in multi-task learning, 2008.

G. Lugosi and N. Et-vayatis, On the Bayes-risk consistency of regularized boosting methods
URL : https://hal.archives-ouvertes.fr/hal-00102140

S. Mallat, A wavelet tour of signal processing The sparse way, With contributions from Gabriel Peyré, 2009.

P. Massart, Concentration inequalities and model selection Lectures from the 33rd Summer School on Probability Theory held in Saint-Flour, Lecture Notes in Mathematics, vol.1896, 2003.

M. Curtis, S. Et-ghosal, and S. , Approximate posterior model probabilities in additive models via the group lasso, 2008.

A. D. Mcquarrie and C. Tsai, Regression and time series model selection, 1998.
DOI : 10.1142/3573

L. Meier, S. Van-de-geer, and P. Et-bühlmann, The group lasso for logistic regression, Journal of the Royal Statistical Society: Series B (Statistical Methodology), vol.68, issue.1, 2006.
DOI : 10.1111/j.1467-9868.2007.00627.x

N. R. Meinshausen and . Lasso, Relaxed Lasso, Computational Statistics & Data Analysis, vol.52, issue.1, pp.374-393, 2007.
DOI : 10.1016/j.csda.2006.12.019

N. Meinshausen and P. Et-bühlmann, High-dimensional graphs and variable selection with the Lasso, The Annals of Statistics, vol.34, issue.3, pp.1436-1462, 2006.
DOI : 10.1214/009053606000000281

N. Meinshausen and P. Et-bühlmann, Stability selection. Manuscript, 2008.

N. Meinshausen and B. Et-yu, Lasso-type recovery of sparse representations for highdimensional data, Ann. Statist, 2006.

Y. Nardi and A. Et-rinaldo, On the asymptotic properties of the group lasso estimator for linear models, Electronic Journal of Statistics, vol.2, issue.0, pp.605-633, 2008.
DOI : 10.1214/08-EJS200

A. Nemirovski, Topics in non-parametric statistics In Lectures on probability theory and statistics (Saint-Flour, Lecture Notes in Math, vol.1738, pp.85-277, 1998.

G. Obozinski, M. J. Wainwright, and M. I. Et-jordan, Union support recovery in highdimensional multivariate regression, 2008.

M. Osborne, B. Presnell, and B. Et-turlach, On the LASSO and its dual, J. Comput. Graph. Statist, vol.9, issue.2, pp.319-337, 2000.

M. R. Osborne, B. Presnell, and B. A. Et-turlach, A new approach to variable selection in least squares problems, IMA Journal of Numerical Analysis, vol.20, issue.3, pp.389-403, 2000.
DOI : 10.1093/imanum/20.3.389

R. Penrose, A generalized inverse for matrices, Mathematical Proceedings of the Cambridge Philosophical Society, vol.11, issue.03, pp.406-413, 1955.
DOI : 10.1093/qmath/2.1.189

B. T. Polyak and A. B. Et-tsybakov, Asymptotic optimality of the C p -test in the projection estimation of a regression, Teor. Veroyatnost. i Primenen, vol.35, issue.2, pp.305-317, 1990.

B. M. Pötscher, Order Estimation in ARMA-Models by Lagrangian Multiplier Tests, The Annals of Statistics, vol.11, issue.3
DOI : 10.1214/aos/1176346253

B. M. Pötscher and H. Et-leeb, On the Distribution of Penalized Maximum Likelihood Estimators: The LASSO, SCAD, and Thresholding, SSRN Electronic Journal, 2007.
DOI : 10.2139/ssrn.1027629

B. M. Pötscher and U. Schneider, On the distribution of the adaptive LASSO estimator, Journal of Statistical Planning and Inference, vol.139, issue.8, 2008.
DOI : 10.1016/j.jspi.2009.01.003

A. Rinaldo, Properties and refinements of the fused lasso, The Annals of Statistics, vol.37, issue.5B, 2008.
DOI : 10.1214/08-AOS665

S. Rosset and J. Et-zhu, Piecewise linear regularized solution paths, The Annals of Statistics, vol.35, issue.3
DOI : 10.1214/009053606000001370

G. Schwartz, Estimating the Dimension of a Model, The Annals of Statistics, vol.6, issue.2, pp.461-464, 1978.
DOI : 10.1214/aos/1176344136

J. Shao, Linear Model Selection by Cross-validation, Journal of the American Statistical Association, vol.39, issue.422, pp.486-494, 1993.
DOI : 10.1080/03610927508827223

J. Shao, An asymptotic theory for linear model selection, Statist. Sinica, vol.7, issue.2, pp.221-264, 1997.

X. Shen and J. Et-ye, Adaptive Model Selection, Journal of the American Statistical Association, vol.97, issue.457, pp.210-221, 2002.
DOI : 10.1198/016214502753479356

R. Shibata, An optimal selection of regression variables, Biometrika, vol.68, issue.1, pp.45-54, 1981.
DOI : 10.1093/biomet/68.1.45

R. Shibata, Selection of the number of regression variables; A minimax choice of generalized FPE, Annals of the Institute of Statistical Mathematics, vol.9, issue.1, pp.459-474, 1986.
DOI : 10.1007/BF02482533

M. Stone, . Corrigendumj, and . Roy, Cross-validatory choice and assessment of statistical predictions, Statist. Soc. Ser. B, vol.36, pp.111-147, 1974.

B. Tarigan and S. A. Et-van-de-geer, Classifiers of support vector machine type with \ell1 complexity regularization, Bernoulli, vol.12, issue.6, pp.1045-1076, 2006.
DOI : 10.3150/bj/1165269150

R. Tibshirani, Regression shrinkage and selection via the lasso, J. Roy. Statist. Soc. Ser. B, vol.58, issue.1, pp.267-288, 1996.

R. Tibshirani, M. Saunders, S. Rosset, J. Zhu, and K. Et-knight, Sparsity and smoothness via the fused lasso, Journal of the Royal Statistical Society: Series B (Statistical Methodology), vol.99, issue.1, pp.91-108, 2005.
DOI : 10.1016/S0140-6736(02)07746-2

J. A. Tropp, Greed is Good: Algorithmic Results for Sparse Approximation, IEEE Transactions on Information Theory, vol.50, issue.10, pp.2231-2242, 2004.
DOI : 10.1109/TIT.2004.834793

A. Tsybakov, Cours de statistiques appliquée, Université Paris, vol.6, p.202

A. Tsybakov, Optimal Rates of Aggregation, Proceedings of 16th Annual Conference on Learning Theory (COLT) and 7th Annual Workshop on Kernel Machines, pp.303-313, 2003.
DOI : 10.1007/978-3-540-45167-9_23

URL : https://hal.archives-ouvertes.fr/hal-00104867

A. B. Tsybakov and S. A. Van-de-geer, Square root penalty: Adaptation to the margin in classification and in edge estimation, The Annals of Statistics, vol.33, issue.3, pp.1203-1224, 2005.
DOI : 10.1214/009053604000001066

URL : https://hal.archives-ouvertes.fr/hal-00101837

B. A. Turlach, On algorithms for solving least squares problems under an ? 1 penalty or an ? 1 constraint, Proceedings of the American Statistical Association. Statistical Computing Section [CD-ROM], pp.2572-2577, 2004.

S. Van-de-geer, The deterministic lasso, 2007.

S. Van-de-geer, High-dimensional generalized linear models and the lasso, The Annals of Statistics, vol.36, issue.2, pp.614-645, 2008.
DOI : 10.1214/009053607000000929

V. Vapnik, The Nature of Statistical Learning Theory, 1998.

V. Vapnik, Statistical learning theory Adaptive and Learning Systems for Signal Processing, Communications, and Control, 1998.

V. Vovk, Asymptotic Optimality of Transductive Confidence Machine, Algorithmic learning theory, pp.336-350
DOI : 10.1007/3-540-36169-3_27

V. Vovk, On-line confidence machines are well-calibrated, The 43rd Annual IEEE Symposium on Foundations of Computer Science, 2002. Proceedings., pp.187-196, 2002.
DOI : 10.1109/SFCS.2002.1181895

V. Vovk, A. Gammerman, and C. Et-saunders, Machine-learning applications of algorithmic randomness, Proceedings of the Sixteenth International Conference on Machine Learning, pp.444-453, 1999.

V. Vovk, A. Gammerman, and G. Et-shafer, Algorithmic learning in a random world, 2005.

V. Vovk, N. Ilia, G. Et-gammerman, and A. , On-line predictive linear regression, The Annals of Statistics, vol.37, issue.3, 2007.
DOI : 10.1214/08-AOS622

URL : http://arxiv.org/abs/0906.3123

M. Wainwright, Sharp thresholds for noisy and high-dimensional recovery of sparsity using l1-constrained quadratic programming, 2006.

H. Wang, G. Li, and G. Et-jiang, Robust Regression Shrinkage and Consistent Variable Selection Through the LAD-Lasso, Journal of Business & Economic Statistics, vol.25, issue.3, pp.347-355, 2007.
DOI : 10.1198/073500106000000251

M. Wegkamp, Model selection in nonparametric regression, The Annals of Statistics, vol.31, issue.1, pp.252-273, 2003.
DOI : 10.1214/aos/1046294464

D. M. Witten and R. Et-tibshirani, Covariance-regularized regression and classification for high dimensional problems, Journal of the Royal Statistical Society: Series B (Statistical Methodology), vol.67, issue.3, 2009.
DOI : 10.1111/j.1467-9868.2009.00699.x

Y. Yang, Can the strengths of AIC and BIC be shared? A conflict between model indentification and regression estimation, Biometrika, vol.92, issue.4, pp.937-950, 2005.
DOI : 10.1093/biomet/92.4.937

Y. Yang, Regression with multiple candidate models : selecting or mixing ? Statist. Sinica, pp.783-809, 2003.

Y. Yang, Aggregating regression procedures to improve performance, Bernoulli, vol.10, issue.1, pp.25-47, 2004.
DOI : 10.3150/bj/1077544602

M. Yuan and Y. Et-lin, Model selection and estimation in regression with grouped variables, Journal of the Royal Statistical Society: Series B (Statistical Methodology), vol.58, issue.1
DOI : 10.1198/016214502753479356

M. Yuan and Y. Et-lin, On the non-negative garrotte estimator, Journal of the Royal Statistical Society: Series B (Statistical Methodology), vol.101, issue.2, pp.143-161, 2007.
DOI : 10.1111/j.1467-9868.2005.00503.x

Y. , A. B. Nazin, A. V. Tsybakov, A. B. Et-vayatis, and N. , Recursive aggregation of estimators by the mirror descent method with averaging, Problemy Peredachi Informatsii, vol.41, issue.4, pp.78-96, 2005.

C. Zhang, Nearly unbiased variable selection under minimax concave penality, 2008.

C. Zhang and J. Huang, The sparsity and bias of the Lasso selection in high-dimensional linear regression, The Annals of Statistics, vol.36, issue.4, pp.1567-1594, 2008.
DOI : 10.1214/07-AOS520

P. Zhao and B. Et-yu, On model selection consistency of Lasso, J. Mach. Learn. Res, vol.7, pp.2541-2563, 2006.

S. Zhou, S. Van-de-geer, and P. Et-bühlmann, Adaptive lasso for high dimensional regression and gaussian graphical modeling, 2009.

H. Zou, The Adaptive Lasso and Its Oracle Properties, Journal of the American Statistical Association, vol.101, issue.476, pp.1418-1429, 2006.
DOI : 10.1198/016214506000000735

H. Zou and T. Et-hastie, Regularization and variable selection via the elastic net, J. R

H. Zou and R. Li, One-step sparse estimates in nonconcave penalized likelihood models

H. Zou, T. Hastie, and R. Et-tibshirani, On the ???degrees of freedom??? of the lasso, The Annals of Statistics, vol.35, issue.5, pp.2173-2192, 2007.
DOI : 10.1214/009053607000000127