Abstract

In the last decade, data generated from different digital devices has posed a remarkable challenge for data representation and analysis. Because of the high-dimensional datasets and the rapid growth of data volume, a lot of challenges have been encountered in various fields such as data mining and data science. Conventional machine learning classifiers are of limited ability to handle the problems of high dimensionality that includes memory limitation, computational cost, and low accuracy performance. Consequently, there is a need to reduce the dimension of datasets by choosing the most significant features that would represent the data efficiently with minimum volume. This study proposes an improved binary version of the equilibrium optimizer algorithm (IBEO) to mitigate features selection problem. Two main enhancements are added to the original equilibrium optimizer (EO) to strengthen its performance. Opposition based learning is the first advancement added to the initialization stage of EO to enhance the diversity of the population in the search space. Local search algorithm is the second advancement added to enhance the exploitation of EO. Wrapper approaches can offer premium solutions. Thus, we used k-nearest neighbour classifier and support vector machine classifiers as the most popular wrapper methods. Moreover, dealing with the problem of over-fitting is an essential task that urges on applying k-fold cross-validation to split each dataset into training and testing data. Comparative tests with different well-known algorithms such as grey wolf optimization, grasshopper optimization, particle swarm optimization, whale optimization, dragonfly, and improved salp swarm algorithms are considered. The proposed algorithm is applied to the most commonly datasets used in the field to validate the performance. Statistical analysis studies demonstrate the effectiveness of the IBEO.

Highlights

  • The technological evolution in many fields such as finance, biomedical, bioinformatics, and telecommunication has produced an exponential volume of pervasive data

  • DATASETS There are 25 datasets taken from https://www.openml.org were used for verifying and evaluating the performance of IBEO compared to other algorithms

  • The classification process is responsible for classifying new incoming instances where the class label is unknown. k nearest neighbor (kNN) and support vector machines (SVMs) are the preferred classifiers in the present study

Read more

Summary

INTRODUCTION

The technological evolution in many fields such as finance, biomedical, bioinformatics, and telecommunication has produced an exponential volume of pervasive data. Sequential forward selection (SFS) is a sequential search strategy that works best when the optimal subset has a limited number of features. The author created a binary version (BMOGW-S) based on a sigmoidal function to solve multi-objective features selection problems with an artificial neural network for classification process [33]. In terms of classification accuracy and the smallest number of selected features, the proposed method outperformed the other compared algorithms. To the best of the author’s information, there are a few studies in the literature for the binary version of EO [83], [84] This has motivated us in this study to propose a new binary version of EO and test its benefit in features selection problems as a binary optimization algorithm.

EQUILIBRIUM OPTIMIZER ALGORITHM
INITIALIZATION
TRANSFORMATION FUNCTION
EXPERIMENTAL RESULTS AND ANALYSIS
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call