Abstract

This study presents a measure-theoretic approach to estimate the upper bound on the total variation of the difference between hypergeometric and binomial distributions using the Kullback-Leibler information divergence. The binomial distribution can be used to find the probabilities associated with the binomial experiments. But if the sample size is large relative to the population size, the experiment may not be binomial, and a binomial distribution is not a good choice to find the probabilities associated with the experiment. The hypergeometric probability distribution is the appropriate probability model to be used when the sample size is large compared to the population size. An upper bound for the total variation in the distance between the hypergeometric and binomial distributions is derived using only the sample and population sizes. This upper bound is used to demonstrate how the hypergeometric distribution uniformly converges to the binomial distribution when the population size increases relative to the sample size.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call