Abstract

In the past decades, the ensemble systems have been shown as an efficient method to increase the accuracy and stability of classification algorithms. However, how to get a valid combination of multiple base-classifiers is still an open question to be solved. In this paper, based on the genetic algorithm, a new self-adaptive stacking ensemble model (called SSEM) is proposed. Different from other ensemble learning classification algorithms, SSEM selectively integrates different base-classifiers, and automatically selects the optimal base-classifier combination and hyper-parameters of base-classifiers via the genetic algorithm. It is noted that all of machine learning methods can be the components of SSEM. In this work, based on two base-classifier selection principles (low complexity of base-classifier and high diversity between different base-classifiers), we select five state-of-art classifiers including Naive Bayes (NB), Extremely Randomized trees (ERT), Logistic, Random Forest (RF) and Classification and Regression Tree (CART) as the base-classifiers of SSEM. To demonstrate the efficiency of SSEM, we have applied it to nine different datasets. Compared with other 11 state-of-art classifiers (NB, ERT, Logistic, RF, CART, Back Propagation Network (BPN), Support Vector Machine (SVM), AdaBoost, Bagging, Convolutional Neural Networks (CNN) and Deep neural network (DNN)), SSEM always performs the best under the five evaluation indexes (Accuracy, Recall, AUC, F1-score and Matthews correlation coefficient (MCC)). Moreover, the significance test result shows that SSEM can achieve highly competitive performance against the other 11 state-of-art classifiers. Altogether, it is evident that SSEM can be a useful framework for classification.

Highlights

  • In recent years, statistical classification has gained massive attention in machine learning [1]–[9]

  • In order to show the performance of the stacking ensemble model (SSEM) method, we compare it with other 11 state-of-art classification algorithms including Naïve Bayes (NB) [51], Extremely Randomized trees (ERT) [52], Logistic [53], Random Forest (RF) [54], Classification and Regression Tree (CART) [55], Back Propagation Network (BPN) [56], support vector machines (SVM) [57], AdaBoost [58], Bagging [59], Convolutional Neural Network (CNN) [60] and Deep neural network (DNN) [61]

  • To show the robustness of our proposed method, we perform the comparison with other classification methods in 9 different datasets

Read more

Summary

INTRODUCTION

Statistical classification has gained massive attention in machine learning [1]–[9]. In the ensemble learning process, the combination of different base-classifiers can provide complementary information concerning unknown examples. The ensemble learning can be roughly divided into three categories: Boosting, Bagging and Stacking. When it comes to Bagging that is computationally intensive, it refers to one of the most updated and effective methods They are adopted for improving unstable estimation and classification schemes. To address the above issues, we propose a novel self-adaptive Stacking ensemble model (called SSEM) to determine optimal base-classifier combination and their corresponding hyperparameters. We design a stacking ensemble learning classification algorithm to integrate base-classifiers with low complexity and high diversity. To overcome the shortcomings of traditional ensemble learning methods, we use the genetic algorithm to automatically select the optimal base-classifier combination, and their hyper-parameters.

RELATED WORK
SSEM ARCHITECTURE
THE OPTIMAL BASE-CLASSIFIER COMBINATION
THE OPTIMAL BASE-CLASSIFIER HYPER-PARAMETER SELECTION
RESULTS
EXPERIMENTAL PLATFORM AND DATA SOURCE
HYPER-PARAMETER SETTINGS IN 9 BENCHMARK DATASETS
SSEM PERFORMS THE BEST IN 9 BENCHMARK DATASETS
THE SIGNIFICANCE TEST OF SSEM
CONCLUSION
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.