Abstract

To make inference about a group of parameters on high-dimensional data, we develop the method of estimator augmentation for the block Lasso, which is defined via the block norm. By augmenting a block Lasso estimator $\hat{\beta}$ with the subgradient $S$ of the block norm evaluated at $\hat{\beta}$, we derive a closed-form density for the joint distribution of $(\hat{\beta},S)$ under a high-dimensional setting. This allows us to draw from an estimated sampling distribution of $\hat{\beta}$, or more generally any function of $(\hat{\beta},S)$, by Monte Carlo algorithms. We demonstrate the application of estimator augmentation in group inference with the group Lasso and a de-biased group Lasso constructed as a function of $(\hat{\beta},S)$. Our numerical results show that importance sampling via estimator augmentation can be orders of magnitude more efficient than parametric bootstrap in estimating tail probabilities for significance tests. This work also brings new insights into the geometry of the sample space and the solution uniqueness of the block Lasso.

Highlights

  • There has been a fast growth of high-dimensional data in many areas, such as genomics and the social sciences

  • To meet the aforementioned challenges in group inference, we develop the method of estimator augmentation for the block lasso

  • Our development of estimator augmentation for the block lasso enables efficient simulation from the sampling distributions of the group lasso and the de-biased group lasso, which is an essential component in practical applications of these inferential approaches

Read more

Summary

Introduction

There has been a fast growth of high-dimensional data in many areas, such as genomics and the social sciences. Statistical inference for high-dimensional models becomes a necessary tool for scientific discoveries from such data. Significance tests have been performed to screen millions of genomic loci for disease markers. These applications have motivated the recent development in high-dimensional statistical inference. There are various other inferential methods for high-dimensional models [7, 8, 15, 16, 24], some of which are reviewed in [4]

Group inference
Contributions of this work
The basic idea
The KKT conditions
Uniqueness
Estimator augmentation
Sample space
A bijective mapping
Joint density
Examples
Applications in statistical inference
Parametric bootstrap
Importance sampling
Group lasso
A de-biased approach
Other applications
Generalizations
A scaled block lasso
Concluding remarks
Auxiliary lemmas
Characterization of solutions
Proof of sufficiency
Results and derivations in Example 1
Derivations in Example 2
Derivations in Example 3

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.