Abstract

Ensembles of classifiers are among the best performing classifiers available in many data mining applications, including the mining of data streams. Rather than training one classifier, multiple classifiers are trained, and their predictions are combined according to a given voting schedule. An important prerequisite for ensembles to be successful is that the individual models are diverse. One way to vastly increase the diversity among the models is to build an heterogeneous ensemble, comprised of fundamentally different model types. However, most ensembles developed specifically for the dynamic data stream setting rely on only one type of base-level classifier, most often Hoeffding Trees. We study the use of heterogeneous ensembles for data streams. We introduce the Online Performance Estimation framework, which dynamically weights the votes of individual classifiers in an ensemble. Using an internal evaluation on recent training data, it measures how well ensemble members performed on this and dynamically updates their weights. Experiments over a wide range of data streams show performance that is competitive with state of the art ensemble techniques, including Online Bagging and Leveraging Bagging, while being significantly faster. All experimental results from this work are easily reproducible and publicly available online.

Highlights

  • Real-time analysis of data streams is a key area of data mining research

  • One way to vastly improve the performance of ensembles is to build heterogeneous ensembles, consisting of models generated by different techniques, rather than homogeneous ensembles, in which all models are generated by the same technique

  • We ran all ensemble techniques on all data streams

Read more

Summary

Introduction

Real-time analysis of data streams is a key area of data mining research. Many real world collected data are streams where observations come in one by one, and algorithms processing these are often subject to time and memory constraints. As data streams are constantly subject to change, the most accurate classifier for a given interval of observations changes frequently, as illustrated by Fig. 1 In their seminal paper, Littlestone and Warmuth (1994) describe a strategy to weight the vote of ensemble members based on their performance on recent observations and prove certain error bounds. This work is of great theoretical value, it needs non-trivial adjustments to be applicable on practical data streams Based on this approach, we propose a way to measure the performance of ensemble members on recent observations and combine their votes. We define Online Performance Estimation, a framework that provides dynamic weighting of the votes of individual ensemble members across the stream Utilising this framework, we introduce a new ensemble technique that combines heterogeneous models.

Related work
Methods
Online performance estimation
Ensemble composition
Experimental setup
Results
Parameter effect
Grace parameter
Number of active classifiers
Conclusions
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call