Abstract

Information criteria such as the Akaike information criterion (AIC) and Bayesian information criterion (BIC) are commonly used for model selection. However, the current theory does not support unconventional data, so naive use of these criteria is not suitable for data with missing values. Imputation, at the core of most alternative methods, is both distorted as well as computationally demanding. We propose a new approach that enables the use of classic well-known information criteria for model selection when there are missing data. We adapt the current theory of information criteria through normalization, accounting for the different sample sizes used for each candidate model (focusing on AIC and BIC). Interestingly, when the sample sizes are different, our theoretical analysis finds that AICj/nj is the proper correction for AICj that we need to optimize (where nj is the sample size available to the jth model) while −(BICj−BICi)/(nj−ni) is the correction of BIC. Furthermore, we find that the computational complexity of normalized information criteria methods is exponentially better than that of imputation methods. In a series of simulation studies, we find that normalized-AIC and normalized-BIC outperform previous methods (i.e., normalized-AIC is more efficient, and normalized BIC includes only important variables, although it tends to exclude some of them in cases of large correlation). We propose three additional methods aimed at increasing the statistical efficiency of normalized-AIC: post-selection imputation, Akaike sub-model averaging, and minimum-variance averaging. The latter succeeds in increasing efficiency further.

Highlights

  • In statistical research and data mining, methods for selecting the “best” model associated with the observed data are of great importance

  • We presented a new perspective on popular information criteria, Akaike information criterion (AIC) and Bayesian information criterion (BIC), and developed a theory that adapts them to address data containing missing values

  • The traditional information criteria are not applicable in cases where data are missing, as they are based on the assumption that all models have the same constant sample size

Read more

Summary

Introduction

In statistical research and data mining, methods for selecting the “best” model associated with the observed data are of great importance. Early work addressing missing data in the context of model selection took this approach and tried to adapt it; it is challenging to combine the results of variable selection (the last stage) across all imputed datasets in a principled framework, and there are a variety of articles on this issue [12,13,14,15,16,17] Another strategy, presented by [12], is to combine multiple imputation and variable selection in a Bayesian framework. The authors of [19] developed the similar method, multiple imputation random lasso, with the main difference being that their method first performs imputation, and only generates bootstrap samples Both sets of results showed that these approaches are suitable for high-dimensional problems.

Normalized Information Criteria
Normalized AIC
Normalized BIC
Statistical Efficiency and Computational Efficiency
Computational Efficiency
Increasing the Statistical Efficiency
Akaike Sub-Model Averaging
Minimal Variance Sub-Model Averaging
Design
Comparing Model Selection Results between AIC Methods
Extensions of Normalized AIC
Comparing Model Selection Results between BIC Methods
Discussion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call