Debiasing Stochastic Gradient Descent to handle missing values - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2020

Debiasing Stochastic Gradient Descent to handle missing values

Résumé

Stochastic gradient algorithm is a key ingredient of many machine learning methods, particularly appropriate for large-scale learning.However, a major caveat of large data is their incompleteness.We propose an averaged stochastic gradient algorithm handling missing values in linear models. This approach has the merit to be free from the need of any data distribution modeling and to account for heterogeneous missing proportion.In both streaming and finite-sample settings, we prove that this algorithm achieves convergence rate of $\mathcal{O}(\frac{1}{n})$ at the iteration $n$, the same as without missing values. We show the convergence behavior and the relevance of the algorithm not only on synthetic data but also on real data sets, including those collected from medical register.
Fichier principal
Vignette du fichier
sgdNA.pdf (813.27 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-02483651 , version 1 (21-02-2020)
hal-02483651 , version 2 (04-06-2020)

Identifiants

Citer

Aude Sportisse, Claire Boyer, Aymeric Dieuleveut, Julie Josse. Debiasing Stochastic Gradient Descent to handle missing values. NeurIPS 2020 - 34th Conference on Neural Information Processing Systems, Dec 2020, Vancouver, Canada. ⟨hal-02483651v2⟩
281 Consultations
362 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More