Abstract

We consider selection of random predictors for a high-dimensional regression problem with a binary response for a general loss function. An important special case is when the binary model is semi-parametric and the response function is misspecified under a parametric model fit. When the true response coincides with a postulated parametric response for a certain value of parameter, we obtain a common framework for parametric inference. Both cases of correct specification and misspecification are covered in this contribution. Variable selection for such a scenario aims at recovering the support of the minimizer of the associated risk with large probability. We propose a two-step selection Screening-Selection (SS) procedure which consists of screening and ordering predictors by Lasso method and then selecting the subset of predictors which minimizes the Generalized Information Criterion for the corresponding nested family of models. We prove consistency of the proposed selection method under conditions that allow for a much larger number of predictors than the number of observations. For the semi-parametric case when distribution of random predictors satisfies linear regressions condition, the true and the estimated parameters are collinear and their common support can be consistently identified. This partly explains robustness of selection procedures to the response function misspecification.

Highlights

  • Consider a random variable ( X, Y ) ∈ R p × {0, 1} and a corresponding response function defined as a posteriori probability q( x ) = P(Y = 1| X = x )

  • We show in the following that variable selection problem can be studied for a general loss function imposing certain analytic properties such as its convexity and Lipschitz property

  • We show for supersets and subsets separately that the probability that the minimum of Generalized Information Criterion (GIC) is not attained at s∗ is exponentially small

Read more

Summary

Introduction

Consider a random variable ( X, Y ) ∈ R p × {0, 1} and a corresponding response function defined as a posteriori probability q( x ) = P(Y = 1| X = x ). One of the main estimation methods of q is a parametric approach for which the response function is assumed to have parametric form q ( x ) = q0 ( β T x ) (1). For some fixed β and known q0 ( x ). If Equation (1) holds, that is the underlying structure is correctly specified, it is known that β = argminb∈ R p − { EX,Y (Y log q0 (b T X ) + (1 − Y ) log(1 − q0 (b T X ))}, Entropy 2020, 22, 153; doi:10.3390/e22020153 (2).

Objectives
Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call