Abstract
In this research work an ensemble of bagging, boosting, rotation forest, decorate and random subspace methods with 5 symbolic sub-classifiers in each one is presented. Then a voting methodology is used for the final prediction. In order to decrease training time, before building the ensemble redundant features were removed using a slight filter feature selection method. A comparison with simple bagging, boosting, rotation forest, decorate and random subspace methods ensembles with 25 symbolic sub-classifiers is performed, as well as other well-known combining methods, on standard benchmark datasets. The proposed technique is shown to be more accurate than other related methods in most cases.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.