Abstract

ABSTRACT Conducting model selection on data gives rise to selection uncertainty which, when ignored, invalidates subsequent classical inference which assumes that the model is given before the analysis and is in all its aspects correctly specified. In selective inference, the randomness induced by selection is dealt with by conditioning confidence intervals and p-values on the subspace of the data which leads to the same model selection as the observed data. The main challenge is the characterization of this selection event. We develop an algorithm for conducting approximate post-selection inference for parameters after model selection events which may not be characterizable as polyhedrons. We apply this on the adaptive lasso, the adaptive elastic net and the group lasso. We conduct experiments on simulated and real data, illustrating that the algorithm can both successfully control the false-positive rate and is computationally efficient.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call