Abstract

Supervised feature selection (FS) for classification aims at finding a more discriminative subset from original features, to facilitate classifier training, improve classification performance and enhance model interpretability. Primary normalized cross-covariance operator (NOCCO) is a nonlinear kernel-based dependency measure between features and labels, including two inverse matrices, and its approximated version (ANOCCO) is simplified via exploiting linear kernel for features, delta kernel for labels and Moore–Penrose inverse. In this paper, we apply NOCCO and ANOCCO to FS task. According to sequential forward selection, a forward NOCCO-based FS algorithm (i.e., FoNOCCO) is directly implemented via various accelerating strategies, but its computational complexity is extremely high. To this end, we propose its fast version via maximizing ANOCCO (i.e., FoANOCCO), where each candidate feature is evaluated efficiently according to column recursive algorithm for Moore–Penrose inverse. Theoretical analysis shows that evaluating a candidate feature needs a time complexity of O(N2) in FoANOCCO and O(N3) in FoNOCCO, where N is the number of instances. On eight small-size high-dimensional benchmark data sets, our detailed experiments illustrate that our two forward FS algorithms are superior to six state-of-the-art FS methods with four baseline classifiers, and FoANOCCO runs 70 times faster than FoNOCCO with a 1.79 % accuracy drop.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.