Abstract

Feature selection is critical in reducing the size of data and improving classifier accuracy by selecting an optimum subset of the overall features. Traditionally, each feature is given a score against a particular category (such as using Mutual Information) and the task of feature selection comes down to choosing the top $k$ ranked features with the best average score across all categories. However, this approach has two major drawbacks. Firstly, the maximum or average score of a feature with a class might not necessarily determine its discriminating strength among samples of other classes. Secondly, most feature selection methods only use the scores to select the discriminating features from the corpus without taking into account the redundancy of information provided by the selected features. In this paper, we propose a new feature ranking score measure called the Discriminative Mutual Information (DMI) score. This score helps to select features that distinguish samples of one category against all other categories. Moreover, Non-Redundant Feature Selection (NRFS) heuristic is also proposed that explicitly takes the problem of feature redundancy into account when selecting the features set. The performance of our approach is investigated and compared with other feature selection techniques on datasets derived from high-dimensional text corpora using multiple classification algorithms. The results show that the proposed method leads to better classification micro-F1 score as compared to other state-of-the-art methods. In particular, the proposed method shows great improvement when the number of selected features are small as well as an overall higher robustness to label noise.

Highlights

  • With the rapid increase in our capacity to generate data, storing and retrieving data efficiently has become increasingly difficult

  • We provide an empirical analysis of several popular feature selection techniques employed in text categorization

  • It is usually not straight forward to compare which feature selection technique is better since different feature subsets might have different inherent characteristics and may be suited for some particular tasks

Read more

Summary

Introduction

With the rapid increase in our capacity to generate data, storing and retrieving data efficiently has become increasingly difficult. Only a subset of features (terms/words in this case) are helpful in discriminating between categories of documents while an overwhelming majority are usually quite generic in nature, connecting phrases, emphasizing a verb or a noun, etc.

Objectives
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call