LassoNet: A Neural Network with Feature Sparsity - Archive ouverte HAL Accéder directement au contenu
Article Dans Une Revue Journal of Machine Learning Research Année : 2021

LassoNet: A Neural Network with Feature Sparsity

Résumé

Much work has been done recently to make neural networks more interpretable, and one approach is to arrange for the network to use only a subset of the available features. In linear models, Lasso (or 1-regularized) regression assigns zero weights to the most irrelevant or redundant features, and is widely used in data science. However the Lasso only applies to linear models. Here we introduce LassoNet, a neural network framework with global feature selection. Our approach achieves feature sparsity by adding a skip (residual) layer and allowing a feature to participate in any hidden layer only if its skip-layer representative is active. Unlike other approaches to feature selection for neural nets, our method uses a modified objective function with constraints, and so integrates feature selection with the parameter learning directly. As a result, it delivers an entire regularization path of solutions with a range of feature sparsity. We apply LassoNet to a number of real-data problems and find that it significantly outperforms state-of-the-art methods for feature selection and regression. LassoNet uses projected proximal gradient descent, and generalizes directly to deep networks. It can be implemented by adding just a few lines of code to a standard neural network.
Fichier principal
Vignette du fichier
1907.12207.pdf (2.2 Mo) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03454793 , version 1 (29-11-2021)

Identifiants

Citer

Ismael Lemhadri, Feng Ruan, Louis Abraham, Robert Tibshirani. LassoNet: A Neural Network with Feature Sparsity. Journal of Machine Learning Research, In press, 22 (127). ⟨hal-03454793⟩
183 Consultations
74 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More