Stochastic Conditional Gradient Method for Composite Convex Minimization - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2019

Stochastic Conditional Gradient Method for Composite Convex Minimization

Résumé

In this paper, we propose the first practical algorithm to minimize stochastic composite optimization problems over compact convex sets. This template allows for affine constraints and therefore covers stochastic semidefinite programs (SDPs), which are vastly applicable in both machine learning and statistics. In this setup, stochastic algorithms with convergence guarantees are either not known or not tractable. We tackle this general problem and propose a convergent, easy to implement and tractable algorithm. We prove $\mathcal{O}(k^{-1/3})$ convergence rate in expectation on the objective residual and $\mathcal{O}(k^{-5/12})$ in expectation on the feasibility gap. These rates are achieved without increasing the batchsize, which can contain a single sample. We present extensive empirical evidence demonstrating the superiority of our algorithm on a broad range of applications including optimization of stochastic SDPs.

Dates et versions

hal-02007612 , version 1 (05-02-2019)

Identifiants

Citer

Francesco Locatello, Alp Yurtsever, Olivier Fercoq, Volkan Cevher. Stochastic Conditional Gradient Method for Composite Convex Minimization. Neural Information Processing Systems, Dec 2019, Vancouver, Canada. ⟨hal-02007612⟩
76 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More