Abstract

With the rapid increase of the data size, it has increasing demands for selecting features by exploiting both labeled and unlabeled data. In this paper, we propose a novel semi-supervised embedded feature selection method. The new method extends the least square regression model by rescaling the regression coefficients in the least square regression with a set of scale factors, which is used for evaluating the importance of features. An iterative algorithm is proposed to optimize the new model. It has been proved that solving the new model is equivalent to solving a sparse model with a flexible and adaptable $\ell _{2,p}$ l 2 , p norm regularization. Moreover, the optimal solution of scale factors provides a theoretical explanation for why we can use $\lbrace \left\Vert \mathbf {w}^{1} \right\Vert _{2},\ldots, \left\Vert \mathbf {w}^{d} \right\Vert _{2}\rbrace$ { w 1 2 , ... , w d 2 } to evaluate the importance of features. Experimental results on eight benchmark data sets show the superior performance of the proposed method.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.