Abstract

Real data is taking on more and more complex structures, raising the necessity for more flexible and parsimonious statistical methodologies. Tensor-variate (or multi-way) structures are a typical example of such kind of data. Unfortunately, real data often present atypical observations that make the traditional normality assumption inadequate. Thus, in this paper, we first introduce two new tensor-variate distributions, both heavy-tailed generalizations of the tensor-variate normal distribution. Then, we use these distributions for model-based clustering via finite mixture models. To introduce parsimony in the models, we use the eigen-decomposition of the components’ scale matrices, obtaining two families of parsimonious tensor-variate mixture models. As a by-product, we also introduce the parsimonious version of tensor-variate normal mixtures. As for parameter estimation, we illustrate variants of the well-known EM algorithm. Since the number of parsimonious models depends on the order of the tensors, we implement strategies intending to shorten the initialization and fitting processes. These procedures are investigated via simulated analyses. Finally, we fitted our parsimonious models to two real datasets having a 4-way and a 5-way structure, respectively.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.