Abstract

In this paper, we consider several binomial mixture models for fitting over-dispersed binary data. The models range from the binomial itself, to the beta-binomial (BB), the Kumaraswamy distributions I and II (KPI \& KPII) as well as the McDonald generalized beta-binomial mixed model (McGBB). The models are applied to five data sets that have received attention in various literature. Because of convergence issues, several optimization methods ranging from the Newton-Raphson to the quasi-Newton optimization algorithms were employed with SAS PROC NLMIXED using the Adaptive Gaussian Quadrature as the integral approximation method within PROC NLMIXED. Our results differ from those presented in Li, Huang and Zhao (2011) for the example data sets in that paper but agree with those presented in Manoj, Wijekoon and Yapa (2013). We also applied these models to the case where we have a $k$ vector of covariates $(x_1, x_2, \ldots, x_k)^{'}$. Our results here suggest that the McGBB performs better than the other models in the GLM framework. All computations in this paper employed PROC NLMIXED in SAS. We present in the appendix a sample of the SAS program employed for implementing the McGBB model for one of the examples.

Highlights

  • The binomial outcome data are widely encountered in many real world applications

  • In this paper we focus the emphasis on the McDonald’s Generalized Beta distribution (Manjor et al, 2015) of the first kind as the mixing distribution and would employ the newly introduced the McDonald Generalized Beta-Binomial distribution(McGBB) Manoj al. (2013)

  • Real world datasets are modeled by using the new McGBB mixture distribution, and it is shown that this model gives better fit than its nested models, namely, the beta-binomial and the Kumaraswamy type II models

Read more

Summary

Introduction

The binomial outcome data are widely encountered in many real world applications. The Binomial distribution often fails to model the binomial outcomes since the variance of the observed binomial outcome data exceeds the nominal Binomial distribution variance, a phenomenon known as over-dispersion. One way of handling overdispersion is modeling the success probability of the Binomial distribution using a continuous distribution defined on the standard unit interval. The resultant general class of univariate discrete distributions is known as the class of Binomial mixture distributions. Real world datasets are modeled by using the new McGBB mixture distribution, and it is shown that this model gives better fit than its nested models, namely, the beta-binomial and the Kumaraswamy type II models. The McGBB and other mixing models consider here model the parameter π of the binomial with continuous distributions defined in the interval (0,1). The beta-binomial The Kumaraswamy binomial I & II binomial models The Macdonald’s generalized beta-binomial distribution http://ijsp.ccsenet.org

The Binomial Model
The Kumaraswamy Distribution
Parameter Estimation
Data Example I
Results
Data Example II-Terrorism in the US
Data Example III-Terrorism Data in Argentina
Data Example IV
Extension to GLM
Conclusions
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call