Abstract
Feature selection (FS) is used to determine the minimal subset of features that are more efficient and could render high accuracy when compared with the whole set of features for the classifiers. Many FS algorithms have been proposed in the literature and it is found that the FS methods which incorporate both relevance and redundant analysis give high accuracy for the classifiers. This paper also provides an alternate novel hybrid-filter-based FS approach for enhancing the predictive accuracy of Naive Bayesian Classifier (NBC) by combining both relevance and redundant analysis. In the relevance analysis, the feature weights relevant to the target are obtained using Bayes Theorem (BT) and Self Information (SI) and the relevant features are selected from the weighted features using Sequential Forward Selection (SFS). But there is a possibility of redundant attributes from the relevant features picked from the relevance analysis and the proposed work removes it by using the information theoretic correlation measure called Symmetric Uncertainty (SU). The efficiency and effectiveness of the proposed algorithm are evaluated by comparing it with the existing methods on eight real-world data sets from the University of California at Irvine UCI machine-learning repository. From the experiments, it has been proved that the proposed work not only reduces the number of features but also enhances the accuracy of NBC.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: International Journal of Data Mining and Emerging Technologies
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.