Abstract

Data from many applied fields exhibit both heavy tail and skewness behavior. For this reason, in the last few decades, there has been a growing interest in exploring parametric classes of skew-symmetrical distributions. A popular approach to model departure from normality consists of modifying a symmetric probability density function in a multiplicative fashion, introducing skewness. An important issue, addressed in this paper, is the introduction of some measures of distance between skewed versions of probability densities and their symmetric baseline. Different measures provide different insights on the departure from symmetric density functions: we analyze and discuss $L_{1}$ distance, $J$-divergence and the concentration function in the normal and Student-$t$ cases. Multiplicative contaminations of distributions can be also considered in a Bayesian framework as a class of priors and the notion of distance is here strongly connected with Bayesian robustness analysis: we use the concentration function to analyze departure from a symmetric baseline prior through multiplicative contamination prior distributions for the location parameter in a Gaussian model.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call