Performance-Complexity Trade-Off in Large Dimensional Statistics - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2020

Performance-Complexity Trade-Off in Large Dimensional Statistics

Résumé

This article introduces a random matrix framework for the analysis of the trade-off between performance and complexity in a class of machine learning algorithms, under a large dimensional data X = [x1,. .. , xn] ∈ R p×n regime. Specifically, we analyze the spectral properties of K B ∈ R n×n , for the kernel random matrix K = 1 p X T X upon which a sparsity mask B ∈ {0, 1} n×n is applied: this reduces the number of Kij to evaluate, thereby reducing complexity, while weakening the power of statistical inference on K, thereby impeding performance. Assuming the data structured as X = Z + √ nµv T for informative vectors µ ∈ R p , v ∈ R n , and white noise Z, we exhibit a phase transition phenomenon below which spectral methods must fail and which is a function of the sparsity structure of B. This finds immediate applications to the fundamental limits of complexity-reduced spectral clustering as well as principal component analysis.
Fichier principal
Vignette du fichier
MLSP_2020_wo_copyright.pdf (732.72 Ko) Télécharger le fichier
Loading...

Dates et versions

hal-02961057 , version 1 (13-10-2020)

Identifiants

Citer

Tayeb Zarrouk, Romain Couillet, Florent Chatelain, Nicolas Le Bihan. Performance-Complexity Trade-Off in Large Dimensional Statistics. MLSP 2020 - IEEE 30th International Workshop on Machine Learning for Signal Processing, Sep 2020, Espoo (virtual), Finland. ⟨10.1109/MLSP49062.2020.9231568⟩. ⟨hal-02961057⟩
143 Consultations
145 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More