Abstract

Imbalanced classification refers to problems in which there are significantly more instances available for some classes than for others. Such scenarios require special attention because traditional classifiers tend to be biased towards the majority class which has a large number of examples. Different strategies, such as re-sampling, have been suggested to improve imbalanced learning. Ensemble methods have also been proven to yield promising results in the presence of class-imbalance. However, most of them only deal with binary imbalanced datasets. In this paper, we propose a re-sampling approach based on belief function theory and ensemble learning for dealing with class imbalance in the multi-class setting. This technique assigns soft evidential labels to each instance. This evidential modeling provides more information about each object's region, which improves the selection of objects in both undersampling and oversampling. Our approach firstly selects ambiguous majority instances for undersampling, then oversamples minority objects through the generation of synthetic examples in borderline regions to better improve minority class borders. Finally, to improve the induced results, the proposed re-sampling approach is incorporated into an evidential classifier-independent fusion-based ensemble. The comparative study against well-known ensemble methods reveals that our method is efficient according to the G-Mean and F1-score measures, independently from the chosen classifier.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call