Abstract

Attributes in datasets are usually not equally significant. Some attributes are unnecessary or redundant. Attribute reduction is an important research issue of rough set theory, which can find minimum subsets of attributes with the same classification effect as the whole dataset by removing unnecessary or redundant attributes. We use Chi-square statistics to evaluate the significance of condition attributes. It can reduce the search space of attribute reduction and improve the speed of attribute reduction. Conditional entropy of relative attributes is adopted as a heuristic function. Two decision table reduction algorithms, forward selection and backward deletion, are proposed to approach the optimal solution. Based on this, an efficient incremental attribute reduction method for dynamically changing datasets is proposed by preserving intermediate variables. The intermediate variable is the observation frequency matrix of joint events of each condition attribute and decision attribute. Experimental results show that the proposed algorithms can improve performance in terms of processing time.

Highlights

  • With the development of Internet and the popularization of computer application, datasets are increasing rapidly in the number of objects and attributes

  • Another method is to reduce data dimensions by deleting unnecessary or redundant attributes, such as attribute reduction based on rough set, Chi-square statistics, clustering, neural network, information entropy, support vector machine and so on [7]–[9]

  • Algorithm 3 Incremental Backward Deletion of Attribute Reduction Algorithms based on Chi-Square Statistics (ICAR)

Read more

Summary

Introduction

With the development of Internet and the popularization of computer application, datasets are increasing rapidly in the number of objects and attributes. Using conditional entropy of related attributes as a heuristic function, two decision table reduction algorithms, forward selection and backward deletion, are proposed to approach the optimal solution. In each iteration of selection process the attributes which increment value of conditional entropy is 0 are deleted as shown in algorithm 1.

Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call