Abstract

Some filter methods stemming from statistics or geometry theory select features individually. Hence they neglect the combination of features and lead to suboptimal subset of features. To address this problem, a joint feature weights learning framework, which automatically determines the optimal size of the feature subset and selects the best features corresponding to a given adjacency graph, is proposed in this paper. In particular, our framework imposes nonnegative and l22-norm constraints on feature weights and iteratively learns feature weights jointly and simultaneously. A new minimization algorithm with proved convergence is also developed to optimize the non-convex objective function. Utilizing this framework as a tool, we propose a new unsupervised feature selection algorithm called Joint Laplacian Feature Weights Learning. Experimental results on five real-world datasets demonstrate the effectiveness of our algorithm.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call