Abstract

Consider the multiple linear regression model $$Y = \beta _1 + \beta _2 x_2 + \cdots + \beta _p x_p + e = {\varvec{x}}^T \varvec{\beta }+ e$$ with sample size n. This paper compares the six shrinkage estimators: forward selection, lasso, partial least squares, principal components regression, lasso variable selection, and ridge regression, with large sample theory and two new prediction intervals that are asymptotically optimal if the estimator $${\hat{\varvec{\beta }}}$$ is a consistent estimator of $$\varvec{\beta }$$ . Few prediction intervals have been developed for $$p>n$$ , and they are not asymptotically optimal. For p fixed, the large sample theory for variable selection estimators like forward selection is new, and the theory shows that lasso variable selection is $$\sqrt{n}$$ consistent under much milder conditions than lasso. This paper also simplifies the proofs of the large sample theory for lasso, ridge regression, and elastic net.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call