Abstract

Asymptotic properties of the generalized information criterion for model selection are examined and new conditions under which this criterion is overfitting, consistent, or underfitting are derived.

Highlights

  • Generalized information criterionConsider a family of probability distributions, f (z; θ), where z ∈ Rd+1 consisting of both response, y, and explanatory variables, x ∈ Rd, and θ ∈ Θ ⊂ Rm is a set of parameters

  • Where α ≥ 0 is the tuning parameter, and κ(Sk) is the model size defined as the number of elements in Sk

  • Assume the true model consists of the variable x3

Read more

Summary

Generalized information criterion

Consider a family of probability distributions, f (z; θ), where z ∈ Rd+1 consisting of both response, y, and explanatory variables, x ∈ Rd, and θ ∈ Θ ⊂ Rm is a set of parameters. Where α ≥ 0 is the tuning parameter, and κ(Sk) is the model size defined as the number of elements in Sk. Let Skn ∈ S be the model selected by gic. We use the terms overfitting and underfitting as defined in the sense of efficiency by [4]. In this framework, model selection is to find the best model that is either a true model or a model closest to the true model. The overfitting implies the model has a larger likelihood value

Asymptotic properties
An example of linear regression
Proof of Theorem
Conclusions
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.