Quantum Expectation-Maximization for Gaussian Mixture Models - Archive ouverte HAL Accéder directement au contenu
Pré-Publication, Document De Travail Année : 2019

Quantum Expectation-Maximization for Gaussian Mixture Models

Résumé

The Expectation-Maximization (EM) algorithm is a fundamental tool in unsupervised machine learning. It is often used as an efficient way to solve Maximum Likelihood (ML) estimation problems, especially for models with latent variables. It is also the algorithm of choice to fit mixture models: generative models that represent unlabelled points originating from $k$ different processes, as samples from $k$ multivariate distributions. In this work we define and use a quantum version of EM to fit a Gaussian Mixture Model. Given quantum access to a dataset of $n$ vectors of dimension $d$, our algorithm has convergence and precision guarantees similar to the classical algorithm, but the runtime is only polylogarithmic in the number of elements in the training set, and is polynomial in other parameters - as the dimension of the feature space, and the number of components in the mixture. We generalize further the algorithm in two directions. First, we show how to fit any mixture model of probability distributions in the exponential family. Then, we show how to use this algorithm to compute the Maximum a Posteriori (MAP) estimate of a mixture model: the Bayesian approach to likelihood estimation problems. We discuss the performance of the algorithm on datasets that are expected to be classified successfully by those algorithms, arguing that on those cases we can give strong guarantees on the runtime.

Dates et versions

hal-02411661 , version 1 (15-12-2019)

Identifiants

Citer

Iordanis Kerenidis, Alessandro Luongo, Anupam Prakash. Quantum Expectation-Maximization for Gaussian Mixture Models. 2019. ⟨hal-02411661⟩
42 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More