AbstractFeature selection is a crucial preprocessing step in data mining and machine learning, enhancing model performance and computational efficiency. This paper investigates the effectiveness of the Side-Blotched Lizard Optimization Algorithm (SBLA) for feature selection by developing six novel variants: Sbla-s1, Sbla-s2, Sbla-s3, Sbla-v1, Sbla-v2, and Sbla-v3, each employing distinct S-shaped or V-shaped transfer functions to convert the continuous search space to a binary format. These variants were rigorously evaluated on nineteen benchmark datasets from the UCI repository, comparing their performance based on average classification accuracy, average number of selected features, and average fitness value. The results demonstrated the superiority of Sbla-s3, achieving an average classification accuracy of 92.8% across all datasets, a mean number of selected features of 20, and an average fitness value of 0.08. Furthermore, Sbla-s3 consistently outperformed six other state-of-the-art metaheuristic algorithms, achieving the highest average accuracy on sixteen out of nineteen datasets. These findings establish Sbla-s3 as a promising and effective approach for feature selection, capable of identifying relevant features while maintaining high classification accuracy, potentially leading to improved model performance in various machine learning applications.