Abstract

Bagging predictors have shown to be effective especially when the learners used to train the base classifiers are weak. In this paper, we argue that for very weak (VW) learners, such as DecisionStump, OneR, and SuperPipes, the base classifiers built from boostrap bags are strongly correlated with each other. As a result, a simple bagging (SB) predictor built on such VW learners has very little improvement compared to a single classifier trained from the same data. Alternatively, we propose a Local Lazy Learning based bagging approach (L3B), where base learners are trained from a small instance subset surrounding each test instance. More specifically, given a test instance x, L3B first discovers x?s k nearest neighbours, and then applies progressive sampling to the selected neighbours to train a set of base classifiers, by using a given VW learner. At the last stage, x is labeled as the most frequently voted class of all base classifiers. Experimental results on 32 real-world datasets, including two high dimensional gene expression datasets, demonstrate that L3B significantly outperforms SB for building accurate classifier ensemble models for VW learners.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.