Abstract
1. Introduction. It is a commonplace observation that the sample mean and sample variance from a normal population (based on a random sample) are stochastically independent. Considerably less prosaic is the converse proposition, first proved in 1936 by R. C. Geary [2] (under superfluous restrictions), to the effect that the independence of these two statistics entails normality of the underlying population. This, plus the theorem that if two linear combinations (nonzero coefficients) of a pair of independent random variables are themselves independent, the variables are normally distributed, which was proved by Kac in 1939 [4], are harbingers of what are today referred to as characterization theorems. An extensive bibliography of such theorems appears in [5]. Most of these results have the format: if such-and-such statistics are independent (alternatively, if the distribution of such-and-such a statistic is thus-and-so), the underlying population is so-and-so. The ensuing theorems belong to this genre but adopt a maximum likelihood posture. The first deals with translation (location) parameter and the latter with scale parameter families of distributions. 2. Preliminaries. Since the results expounded here concern maximum likelihood estimators, it would seem appropriate to say a few words concerning these. It is somewhat surprising that major treatises on mathematical statistics and estimation do not define maximum likelihood estimators per se but merely a maximum likelihood estimate. (Pitman's terminological demarcation between these notions will be made explicit shortly.) The definitions of [8], [9] are closest in spirit tQ that given here. In order to pave the way for a discussion of these questions, let F(x; 0), - < X < oo, 0 E Q C R' denote a one parameter family of probability distributions on the real line R' with spectra So . Define S = U0eaSo and S = S X S X ... X S, the n-fold cartesian product of S with itself.' If, for each 0 E Q, F(x; 0) is absolutely continuous, designate its probability density function (p.d.f.) by f(x; 0); if, for each 0, F(x; 0) is a step function, the same notation f(x; 0) will be used to specify the so-called discrete p.d.f., that is, the mass function of
Published Version (Free)
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have