Abstract

The Dirichlet process mixture of normal densities has been successfully used as a prior for Bayesian density estimation for independent and identically distributed (i.i.d.) observations. A Markov model, which generalizes the i.i.d. set up, may be thought of as a suitable framework for observations arising over time. The predictive density of the future observation is then given by the posterior expectation of the transition density given the observations. We consider a Dirichlet process mixture prior for the transition density and study posterior consistency. Like the i.i.d. case, posterior consistency is obtained if the Kullback–Leibler neighborhoods of the true transition density receive positive prior probabilities and uniformly exponentially consistent tests exist for testing the true density against the complement of its neighborhoods. We show that under reasonable conditions, the Kullback–Leibler property holds for the Dirichlet mixture prior. For certain topologies on the space of transition densities, we show consistency holds under appropriate conditions by constructing the required tests. This approach, however, may not always lead to the best possible results. By modifying a recent approach of Walker [2004. New approaches to Bayesian consistency. Ann. Statist. 32, 2028–2043] for the i.i.d. case, we also show that better conditions for consistency can be given for certain weaker topologies.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call