Abstract

Ensemble learning have emerged as a useful machine learning technique, which is based on the idea that combining the output of multiple models instead of using a single model. This practice, known as “diversity”, and it usually enhances the performance. On other hand, ensemble feature selection method is based on the same idea, where multiple feature subsets are combined to select an optimal subset of features. Learning methods have difficulties with the dimensionality curse that impact the performance and increase the time exponentially. To overcome this issue, we propose a parallel heterogeneous ensemble feature selection based on three well-regarded algorithms: genetic algorithm, particle swarm optimizer, and grey wolf optimizer. The proposed approach is based on four phases; namely, distribution phase, parallel ensemble feature selection phase, combining and aggregation phase, and testing phase. Three implementations of the proposed approach are presented: a sequential approach running on the central processing unit (CPU), a parallel approach running on multi-core CPU, and a parallel approach running on multi-core CPU with graphics processing units (GPU). To assess the performance of the proposed approach twenty-one large datasets were used. The results show that the proposed parallel approach improved the performance in terms of the prediction results and running time.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.