Risk Bounds for Learning Multiple Components with Permutation-Invariant Losses - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2020

Risk Bounds for Learning Multiple Components with Permutation-Invariant Losses

Fabien Lauer

Résumé

This paper proposes a simple approach to derive efficient error bounds for learning multiple components with sparsity-inducing regularization. We show that for such regularization schemes, known decompositions of the Rademacher complexity over the components can be used in a more efficient manner to result in tighter bounds without too much effort. We give examples of application to switching regression and center-based clustering/vector quantization. Then, the complete workflow is illustrated on the problem of subspace clustering, for which decomposition results were not previously available. For all these problems, the proposed approach yields risk bounds with mild dependencies on the number of components and completely removes this dependence for nonconvex regularization schemes that could not be handled by previous methods.
Fichier principal
Vignette du fichier
Lauer_permutatioinvariance_HAL.pdf (169.95 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-02100779 , version 1 (16-04-2019)
hal-02100779 , version 2 (23-01-2020)

Identifiants

Citer

Fabien Lauer. Risk Bounds for Learning Multiple Components with Permutation-Invariant Losses. 23rd International Conference on Artificial Intelligence and Statistics (AISTATS), 2020, Palermo, Italy. ⟨hal-02100779v2⟩
76 Consultations
57 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More