Abstract

Mixture distributions arise in many parametric and non-parametric settings—for example, in Gaussian mixture models and in non-parametric estimation. It is often necessary to compute the entropy of a mixture, but, in most cases, this quantity has no closed-form expression, making some form of approximation necessary. We propose a family of estimators based on a pairwise distance function between mixture components, and show that this estimator class has many attractive properties. For many distributions of interest, the proposed estimators are efficient to compute, differentiable in the mixture parameters, and become exact when the mixture components are clustered. We prove this family includes lower and upper bounds on the mixture entropy. The Chernoff α -divergence gives a lower bound when chosen as the distance function, with the Bhattacharyaa distance providing the tightest lower bound for components that are symmetric and members of a location family. The Kullback–Leibler divergence gives an upper bound when used as the distance function. We provide closed-form expressions of these bounds for mixtures of Gaussians, and discuss their applications to the estimation of mutual information. We then demonstrate that our bounds are significantly tighter than well-known existing bounds using numeric simulations. This estimator class is very useful in optimization problems involving maximization/minimization of entropy and mutual information, such as MaxEnt and rate distortion problems.

Highlights

  • A mixture distribution is a probability distribution whose density function is a weighted sum of individual densities

  • It is often necessary to compute the differential entropy [2] of a random variable with a mixture distribution, which is a measure of the inherent uncertainty in the outcome of the random variable

  • Two thousand samples were used for each Monte Carlo (MC) estimate for the mixtures of Gaussians, and 5000 samples were used for the mixtures of uniform distributions

Read more

Summary

Introduction

A mixture distribution is a probability distribution whose density function is a weighted sum of individual densities. We introduce a novel family of estimators for the mixture entropy Each member of this family is defined via a pairwise-distance function between component densities. The estimators in this family have several attractive properties They are computationally efficient, as long as the pairwise-distance function and the entropy of each component distribution are easy to compute. We consider both mixtures of Gaussians and mixtures of uniform distributions

Background and Definitions
Overview
Lower Bound
Upper Bound
Exact Estimation in the “Clustered” Case
Gaussian Mixtures
Estimating Mutual Information
Numerical Results
Mixture of Gaussians
Mixture of Uniforms
Discussion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call