Abstract

Feature selection (FS) is the activity of defining the most contributing feature subset among all used features to improve the superiority of datasets with a large number of dimensions by selecting significant features and eliminating redundant and irrelevant ones. Therefore, this process can be seen as an optimization process. The primary goals of feature selection are to decrease the number of dimensions and enhance classification accuracy in many domains, such as text classification, large-scale data analysis, and pattern recognition. Several metaheuristics, such as the Genghis Khan Shark Optimizer Algorithm (GKSO), can assist in optimizing the FS issue. However, these methods tend to converge towards local solutions with a low convergence rate. In order to address this issue in GKSO, a more refined version called I-GKSO is implemented. The I-GKSO suggested introducing a new approach to modify solutions that have low fitness values. In addition, it employs the Enhanced Solution Quality (ESQ) strategy to enhance the exploration phase. It utilizes Quasi-opposite-based learning (QOBL) to enhance the best solution obtained and, consequently, the entire population. The algorithm presented aims to solve the FS problem and has been assessed using benchmark optimization problems from the CEC’2017 and CEC’2022. To assess the efficacy of the I-GKSO, it has been subjected to comparisons with multiple different algorithms. The trials conducted using FS datasets yield a quantitative consideration of the I-GKSO's capacity to attain the most optimal subset of features. Furthermore, Wilcoxon and Friedman's non-parametric tests were accomplished to support the performance of the proposed method.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call