Abstract
Dirichlet process mixture (DPM) models provide flexible modeling for distributions of data as an infinite mixture of distributions from a chosen collection. Specifying priors for these models in individual data contexts can be challenging. In this paper, we introduce a scheme which requires the investigator to specify only simple scaling information. This is used to transform the data to a fixed scale on which a low information prior is constructed. Samples from the posterior with the rescaled data are transformed back for inference on the original scale. The low information prior is selected to provide a wide variety of components for the DPM to generate flexible distributions for the data on the fixed scale. The method can be applied to all DPM models with kernel functions closed under a suitable scaling transformation. Construction of the low information prior, however, is kernel dependent. Using DPM-of-Gaussians and DPM-of-Weibulls models as examples, we show that the method provides accurate estimates of a diverse collection of distributions that includes skewed, multimodal, and highly dispersed members. With the recommended priors, repeated data simulations show performance comparable to that of standard empirical estimates. Finally, we show weak convergence of posteriors with the proposed priors for both kernels considered.
Highlights
The Dirichlet process mixture (DPM) model was first proposed by Lo (1984)
The choice of kernel density f (·|θ) determines the mixture components to use in a DPM; for example, if f (·|θ) is a normal kernel, this DPM is a mixture of Gaussians
We offer a technique for an omnibus low information prior specification that can handle data of various scales in a mixture-of-Gaussians model and a mixture-of-Weibulls model
Summary
The Dirichlet process mixture (DPM) model was first proposed by Lo (1984). The marginal distribution of a DPM is a convolution of a kernel density function and a Dirichlet process, g(y) = f (y|G)DP (dG). This model uses the Dirichlet process (DP). Of Ferguson (1973) effectively to estimate density functions even though the DP almost surely generates discrete distributions. Each observation yi arises from a density function f (·|θi) with corresponding parameter θi, which in turn arises from a discrete distribution G. The distribution G is LIO Priors for Dirichlet Process Mixture Models randomly generated from a DP with baseline distribution G0 and concentration parameter ν. The choice of kernel density f (·|θ) determines the mixture components to use in a DPM; for example, if f (·|θ) is a normal kernel, this DPM is a mixture of Gaussians
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.