Abstract
Feature selection has become a mandatory step in several data exploration and Machine Learning applications since data quality can have a strong impact in the performance of machine learning models. Many feature selection strategies have been developed in the past decades, using different criteria to select the most relevant features. The use of dynamic feature selection, however, has showed that the use of multiple simultaneously criteria to determine the best attribute subset for similar instances can deliver encouraging results. In this context, this paper proposes to analyze the performance of a pareto-front based dynamic feature selection (PF-DFS) method under data noise scenarios. In order to do this, we intentionally added noise in 15 datasets and evaluated the PF-DFS performance in order to measure its stability under two different data noise scenarios. The obtained results are compared to some state-of-the-art algorithms and show that, in terms of accuracy, the PF-DFS method is more robust to the other methods for the majority of the analyzed scenarios.
Published Version
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have