Non-asymptotic oracle inequalities for the Lasso in high-dimensional mixture of experts - Archive ouverte HAL Accéder directement au contenu
Pré-Publication, Document De Travail Année : 2020

Non-asymptotic oracle inequalities for the Lasso in high-dimensional mixture of experts

Résumé

Mixture of experts (MoE) has a well-principled finite mixture model construction for prediction, allowing the gating network (mixture weights) to learn from the predictors (explanatory variables) together with the experts' network (mixture component densities). We investigate the estimation properties of MoEs in a high-dimensional setting, where the number of predictors is much larger than the sample size, for which the literature lacks computational and especially theoretical results. We consider the class of finite MoE models with softmax gating functions and Gaussian regression experts, and focus on the theoretical properties of their l1-regularized estimation via the Lasso. We provide a lower bound on the regularization parameter of the Lasso penalty that ensures an l1-oracle inequality is satisfied by the Lasso estimator according to the Kullback--Leibler loss. We further state an l1-ball oracle inequality for the l1-penalized maximum likelihood estimator from the model selection.

Dates et versions

hal-02957892 , version 1 (05-10-2020)

Identifiants

Citer

Trungtin Nguyen, Hien D Nguyen, Faicel Chamroukhi, Geoffrey J Mclachlan. Non-asymptotic oracle inequalities for the Lasso in high-dimensional mixture of experts. 2020. ⟨hal-02957892⟩
68 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More