Abstract

One problem in utilizing ensemble feature selection models is machine learning is the fact that there is no guarantee that an ensemble model will improve machine learning classification performance. This implies that different ensemble models have different success probability, i.e. have different probability in improving the performance of machine learning. This paper introduces the concept of success probability for heterogeneous ensemble models and stated the definitions, notations, and algorithms necessary to the mathematical formulation and computation of the success probability. To show how the theory applied, we create an ensemble filter feature selection model that uses four filter feature selection algorithms (Correlation, Gain Ratio, Info Gain, and One R) as base filters and the Max as a combination method. The experimental results showed that the success probability of the developed ensemble filter model using a set of 9 machine learning algorithms is found to be 0.58.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call