Abstract

A New Heuristic for generating height balanced compact decision trees with reduced path length has been proposed in this paper. The proposed node split measure (MGI) modifies attribute selection process of gini index by giving importance to almost equal sized sub partitions and at same time counts frequently occurring target classes in those sub partitions. This results in less skewed decision trees, while maintaining similar accuracy as provided by gini index. As a part of data pre processing step, MGI gives a new approach to handle missing values and also provides initial data set size reduction. The performance evaluation is done on few sample data sets taken from UCI Machine learning repository. The experimental evaluation of MGI node split measure is done with gini index which shows that the proposed split measure produces height balanced decision tree of reduced height, which in turn reduces antecedents in the rule generated.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call