Quantum Expectation-Maximization for Gaussian Mixture Models
August 19, 2019 Β· Declared Dead Β· π International Conference on Machine Learning
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Iordanis Kerenidis, Alessandro Luongo, Anupam Prakash
arXiv ID
1908.06657
Category
quant-ph: Quantum Computing
Cross-listed
cs.DS,
cs.LG,
stat.ML
Citations
28
Venue
International Conference on Machine Learning
Last Checked
2 months ago
Abstract
The Expectation-Maximization (EM) algorithm is a fundamental tool in unsupervised machine learning. It is often used as an efficient way to solve Maximum Likelihood (ML) estimation problems, especially for models with latent variables. It is also the algorithm of choice to fit mixture models: generative models that represent unlabelled points originating from $k$ different processes, as samples from $k$ multivariate distributions. In this work we define and use a quantum version of EM to fit a Gaussian Mixture Model. Given quantum access to a dataset of $n$ vectors of dimension $d$, our algorithm has convergence and precision guarantees similar to the classical algorithm, but the runtime is only polylogarithmic in the number of elements in the training set, and is polynomial in other parameters - as the dimension of the feature space, and the number of components in the mixture. We generalize further the algorithm in two directions. First, we show how to fit any mixture model of probability distributions in the exponential family. Then, we show how to use this algorithm to compute the Maximum a Posteriori (MAP) estimate of a mixture model: the Bayesian approach to likelihood estimation problems. We discuss the performance of the algorithm on a dataset that is expected to be classified successfully by this algorithm, arguing that on those cases we can give strong guarantees on the runtime.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β Quantum Computing
R.I.P.
π»
Ghosted
R.I.P.
π»
Ghosted
The power of quantum neural networks
R.I.P.
π»
Ghosted
Power of data in quantum machine learning
R.I.P.
π»
Ghosted
Quantum machine learning: a classical perspective
R.I.P.
π»
Ghosted
Noise-Adaptive Compiler Mappings for Noisy Intermediate-Scale Quantum Computers
R.I.P.
π»
Ghosted
ProjectQ: An Open Source Software Framework for Quantum Computing
Died the same way β π» Ghosted
R.I.P.
π»
Ghosted
Language Models are Few-Shot Learners
R.I.P.
π»
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
π»
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
π»
Ghosted