Abstract

We propose a greedy variational method for decomposing a non-negative multivariate signal as a weighted sum of Gaussians, which, borrowing the terminology from statistics, we refer to as a Gaussian mixture model. Notably, our method has the following features: (1) It accepts multivariate signals, i.e., sampled multivariate functions, histograms, time series, images, etc., as input. (2) The method can handle general (i.e., ellipsoidal) Gaussians. (3) No prior assumption on the number of mixture components is needed. To the best of our knowledge, no previous method for Gaussian mixture model decomposition simultaneously enjoys all these features. We also prove an upper bound, which cannot be improved by a global constant, for the distance from any mode of a Gaussian mixture model to the set of corresponding means. For mixtures of spherical Gaussians with common variance sigma ^2, the bound takes the simple form sqrt{n}sigma . We evaluate our method on one- and two-dimensional signals. Finally, we discuss the relation between clustering and signal decomposition, and compare our method to the baseline expectation maximization algorithm.

Highlights

  • Mixtures of Gaussians are often used in clustering to fit a probability distribution to some given sample points

  • In each iteration of our algorithm, a new Gaussian is added to the Gaussian mixture model (GMM) by a procedure that corresponds to one iteration of a continuously parameterized version of matching pursuit (MP), c.f. [10]: The starting guess x0 for the mean vector is defined to be a global maximum of r, which is a smoothed version of the current residual r

  • In the first two lemmas we compute some integrals over the n-sphere of certain polynomials, and in the third lemma we provide expressions for the gradient and Hessian of a GMM

Read more

Summary

Introduction

Mixtures of Gaussians are often used in clustering to fit a probability distribution to some given sample points. In this work we are concerned with the related problem of approximating a non-negative but otherwise arbitrary signal by a sparse linear combination of potentially anisotropic Gaussians. Methods for sparse decomposition of multivariate signals as Gaussian mixture models (GMMs) may be considered in two classes. We complement our algorithm with a theorem (Theorem 1) that upper bounds the distance from a local maximum of a GMM to the set of mean vectors. This provides theoretical support for our initialization of each new mean vector.

Problem statement
Proposed method
Main numerical results
Clustering and comparison to expectation maximization
Clustering via signal decomposition
Signal decomposition via clustering
Theoretical considerations
Location of modes of Gaussian mixtures
Function approximation by Gaussian mixtures
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call