Abstract

In this paper, we consider nonparametric multidimensional finite mixture models and we are interested in the semiparametric estimation of the population weights. Here, the i.i.d. observations are assumed to have at least three components which are independent given the population. We approximate the semiparametric model by projecting the conditional distributions on step functions associated to some partition. Our first main result is that if we refine the partition slowly enough, the associated sequence of maximum likelihood estimators of the weights is asymptotically efficient, and the posterior distribution of the weights, when using a Bayesian procedure, satisfies a semiparametric Bernstein von Mises theorem. We then propose a cross-validation like procedure to select the partition in a finite horizon. Our second main result is that the proposed procedure satisfies an oracle inequality. Numerical experiments on simulated data illustrate our theoretical results.

Highlights

  • We consider in this paper multidimensional mixture models that describe the probability distribution of a random vector X with at least three coordinates

  • We prove in Proposition 2 that when the partition is refined, the Fisher information associated to this partition increases. This leads to our main general result presented in Theorem 1, Section 2.3: it is possible to let the approximation parametric models grow with the sample size so that the sequence of maximum likelihood estimators are asymptotically efficient in the semiparametric model and so that a semiparametric Bernstein-von Mises Theorem holds

  • 3 Model selection In Theorem 1, we prove the existence of some increasing partition leading to efficiency

Read more

Summary

Introduction

We consider in this paper multidimensional mixture models that describe the probability distribution of a random vector X with at least three coordinates. In the particular case of the semiparametric mixtures, this is of great interest, since the construction of a confidence region is not necessarily trivial This is our first main result which is stated in Theorem 1: by considering partitions refined slowly enough when the number of observations increases, we can derive efficient estimation procedures for the parameter of interest θ and in the Bayesian approach for a marginal posterior distribution on θ which satisfies the renowned Bernstein-von Mises property. We prove in Proposition 2 that when the partition is refined, the Fisher information associated to this partition increases This leads to our main general result presented in Theorem 1, Section 2.3: it is possible to let the approximation parametric models grow with the sample size so that the sequence of maximum likelihood estimators are asymptotically efficient in the semiparametric model and so that a semiparametric Bernstein-von Mises Theorem holds.

Model and notations
Efficient influence functions and efficient Fisher informations
General result
Model selection
Behaviour of the MLE as M increases
Criterion for model selection
On the estimation of the risk and the selection of M
Tk when
On the choice of M0
Conclusion and discussion
On the structural assumptions on the model
Extensions to Hidden Markov models
Proof of Proposition 1
Proof of Proposition 2
Proof of Lemma 1
Proof of Proposition 3
Proof of Corollary 1
Proof of Theorem 2
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.