Abstract

Bagging, as a commonly-used class imbalance learning method, combines resampling techniques with ensemble learning to provide a strong classifier with high generalization for a skewed dataset. However, integrating different numbers of base classifiers may obtain the same classification performance, called multi-modality. To seek the most compact ensemble structure with the highest accuracy, a dual evolutionary bagging framework composed of inner and outer ensemble models is proposed. In inner ensemble model, three sub-classifiers are built by SVM, MLP and DT, respectively, with the purpose of enhancing the diversity among them. For each sub-dataset, a classifier with the best performance is selected as a base classifier of outer ensemble model. Following that, all optimal combinations of base classifiers is found by a multi-modal genetic algorithm with a niche strategy in terms of their average G-mean. A combination that aggregates the smallest number of base classifiers by the weighted sum forms the final ensemble structure. Experimental results on 40 KEEL benchmark datasets and a practical one of coal burst show that dual ensemble framework proposed in the paper provides the simplest ensemble structure with the best classification accuracy for imbalance datasets and outperforms the state-of-the-art ensemble learning methods.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.