Abstract

Penalized regression provides an automated approach to preform simultaneous variable selection and parameter estimation and is a popular method to analyze high-dimensional data. Since the conception of the LASSO in the mid-to-late 1990s, extensive research has been done to improve penalized regression. The LASSO, and several of its variations, performs penalization symmetrically around zero. Thus, variables with the same magnitude are shrunk the same regardless of the direction of effect. To the best of our knowledge, sign-based shrinkage, preferential shrinkage based on the sign of the coefficients, has yet to be explored under the LASSO framework. We propose a generalization to the LASSO, asymmetric LASSO, that performs sign-based shrinkage. Our method is motivated by placing an asymmetric Laplace prior on the regression coefficients, rather than a symmetric Laplace prior. This corresponds to an asymmetric ℓ 1 penalty under the penalized regression framework. In doing so, preferential shrinkage can be performed through an auxiliary tuning parameter that controls the degree of asymmetry. Our numerical studies indicate that the asymmetric LASSO performs better than the LASSO when effect sizes are sign skewed. Furthermore, in the presence of positively-skewed effects, the asymmetric LASSO is comparable to the non-negative LASSO without the need to place an a priori constraint on the effect estimates and outperforms the non-negative LASSO when negative effects are also present in the model. A real data example using the breast cancer gene expression data from The Cancer Genome Atlas is also provided, where the asymmetric LASSO identifies two potentially novel gene expressions that are associated with BRCA1 with a minor improvement in prediction performance over the LASSO and non-negative LASSO.

Highlights

  • Recent developments in data acquisition, collection, and storage have allowed researchers to obtain a large number of potential predictors in order to avoid missing important factors that may be associated with the outcome of interest

  • While estimation is focused under a penalized regression framework, we provide a Bayesian interpretation that motivates the use of the asymmetric 1 penalty

  • We develop a generalization to LASSO penalization that asymmetrically penalizes coefficients based on sign

Read more

Summary

Introduction

Recent developments in data acquisition, collection, and storage have allowed researchers to obtain a large number of potential predictors in order to avoid missing important factors that may be associated with the outcome of interest. This is often the case in genomic studies, where the number of predictors collected is often larger than the sample size. Penalized regression methods accomplish this by shrinking the regression coefficients toward zero while setting some coefficients equal to zero. These methods estimate a sparse vector of regression coefficients by minimizing an objective function that is composed of both a loss function and a penalty function.

Objectives
Methods
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.