Abstract

Typically, when considering multiple classifiers, researchers assume that they are independent. Under this assumption estimates for the performance of the fused classifiers are easier to obtain and quantify mathematically. But, in fact, classifiers may be correlated, thus, the performance of the fused classifiers will be over-estimated. This paper will address the issue of the dependence between the classifiers to be fused. Specifically, we will assume a level of dependence between two classifiers for a given fusion rule and produce a formula to quantify the performance of this newly fused classifier. The performance of the fused classifiers will then be evaluated via the Receiver Operating Characteristic (ROC) curve. A classifier typically relies on parameters that may vary over a given range. Thus, the probability of true and false positives can be computed over this range of values. The graph of these probabilities over this range then produces the ROC curve. The probability of true positive and false positive from the fused classifiers are developed according to various decision rules. Examples of dependent fused classifiers will be given for various levels of dependency and multiple decision rules.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call