Minimax risks for sparse regressions: Ultra-high-dimensional phenomenons - Archive ouverte HAL Accéder directement au contenu
Pré-Publication, Document De Travail Année : 2010

Minimax risks for sparse regressions: Ultra-high-dimensional phenomenons

Résumé

Consider the standard Gaussian linear regression model $Y= X\theta+\epsilon$, where $Y\in R^n$ is a response vector and $X \in R^{n\times p}$ is a design matrix. Numerous work have been devoted to building efficient estimators of $\theta$ when $p$ is much larger than $n$. In such a situation, a classical approach amounts to assume that $\theta$ is approximately sparse. This paper studies the minimax risks of estimation and testing over classes of $k$-sparse vectors $\theta$. These bounds shed light on the limitations due to high-dimensionality. The results encompass the problem of prediction (estimation of ${\bf X}\theta$), the inverse problem (estimation of $\theta$) and linear testing (testing ${\bf X}\theta=0$). Interestingly, an elbow effect occurs when the number of variables $k\log(p/k)$ becomes large compared to $n$. Indeed, the minimax risks and hypothesis separation distances blow up in this ultra-high dimensional setting. We also prove that even dimension reduction techniques cannot provide satisfying results in a ultra-high dimensional setting. Moreover, we compute the minimax risks when the variance of the noise is unknown. The knowledge of this variance is shown to play a significant role in the optimal rates of estimation and testing.
Fichier principal
Vignette du fichier
ultra-high.pdf (550.2 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-00508339 , version 1 (03-08-2010)
hal-00508339 , version 2 (20-09-2010)
hal-00508339 , version 3 (23-01-2012)

Identifiants

Citer

Nicolas Verzelen. Minimax risks for sparse regressions: Ultra-high-dimensional phenomenons. 2010. ⟨hal-00508339v2⟩
145 Consultations
226 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More