Abstract
Abstract Using both frequentist and Bayesian techniques, predicting densities are derived for future observations from a multivariate linear model with matrix normal error terms. All the candidates belong to a general location-scale family of predicting densities. An analytic comparison is undertaken, using a Kullback—Leibler loss, by citing an optimal member of a subclass including most of these predicting densities as members. The subclass is based on an invariant Student-t random matrix, and the optimal member is the Bayesian predictive density corresponding to a Jeffreys noninformative prior. Information-based numerical comparisons illustrate the nature of the dominance.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.