Abstract

Feature selection is an important pre-processing step in machine learning and data mining tasks, which improves the performance of the learning models by removing redundant and irrelevant features. Many feature selection algorithms have been widely studied, including greedy and random search approaches, to find a subset of the most important features for fulfilling a particular task (i.e., classification and regression). As a powerful swarm-based meta-heuristic method, particle swarm optimization (PSO) is reported to be suitable for optimization problems with continuous search space. However, the traditional PSO has rarely been applied to feature selection as a discrete space search problem. In this paper, a novel feature selection algorithm based on PSO with learning memory (PSO-LM) is proposed. The goal of the learning memory strategy is designed to inherit much more useful knowledge from those individuals who have higher fitness and offer faster progress, and the genetic operation is used to balance the local exploitation and the global exploration of the algorithm. Moreover, the $k$ -nearest neighbor method is used as a classifier to evaluate the classification accuracy of a particle. The proposed method has been evaluated on some international standard data sets, and the results demonstrated its superiority compared with those wrapper-based feature selection methods.

Highlights

  • In machine learning and data mining, feature selection as a combinatorial optimization problem is an important pre-processing step to obtain the less correlated and distinct feature subset from the original feature set

  • One of the two memories is used for recording those individuals who have higher fitness, while the other is used for saving those individuals who offer faster progress

  • A potential exemplar, based on two individuals who have been selected from the two memories individually, will be generated for a particle

Read more

Summary

INTRODUCTION

In machine learning and data mining, feature selection as a combinatorial optimization problem is an important pre-processing step to obtain the less correlated and distinct feature subset from the original feature set. Since the wrapper approach evaluates a feature subset using learning algorithms with the evaluation in search process, it often obtains better results than the filter one. The classification error rate and number of features are two main capability parameters in search processes, VOLUME 7, 2019 most of PSO-based FS methods only emphasize minimizing the former other than both of them. Inspired by CSO, a new PSO variant, called particle swarm optimization with learning memory (PSO-LM), is presented for solving feature subset selection problem in this paper. Instead of learning from the global and personal best position, a particle in PSO-LM learns from all personal best individuals of the current generation and the exemplar generated above In this way, different helpful knowledge, which comes from those individuals who have higher fitness and offer faster progress, has been utilized effectively.

BACKGROUND
13: End For 14
Findings
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call