Abstract

For evaluating the classification model of an information system, a proper measure is usually needed to determine if the model is appropriate for dealing with the specific domain task. Though many performance measures have been proposed, few measures were specially defined for multi-class problems, which tend to be more complicated than two-class problems, especially in addressing the issue of class discrimination power. Confusion entropy was proposed for evaluating classifiers in the multi-class case. Nevertheless, it makes no use of the probabilities of samples classified into different classes. In this paper, we propose to calculate confusion entropy based on a probabilistic confusion matrix. Besides inheriting the merit of measuring if a classifier can classify with high accuracy and class discrimination power, probabilistic confusion entropy also tends to measure if samples are classified into true classes and separated from others with high probabilities. Analysis and experimental comparisons show the feasibility of the simply improved measure and demonstrate that the measure does not stand or fall over the classifiers on different datasets in comparison with the compared measures.

Highlights

  • Classifier evaluation is one of the fundamental issues in the machine learning and pattern recognition societies, especially when a new classification method is introduced and compared with other possibleEntropy 2013, 15 candidates

  • Besides taking into account both the classification accuracy and class discrimination power of classifiers, probabilistic confusion entropy tries to measure if samples are classified into true classes with high probabilities and into other classes unevenly with low probabilities

  • The measure of confusion entropy is improved for evaluating classification models of information systems

Read more

Summary

Introduction

Classifier evaluation is one of the fundamental issues in the machine learning and pattern recognition societies, especially when a new classification method is introduced and compared with other possible. Some researchers strongly recommended to evaluate classification models by a graphical, multi-objective analysis method, such as the ROC (Receiver operating characteristic) analysis [1,2], instead of by scalar performance measures Though they were proposed for evaluating classifiers from different aspects, most of the measures and analysis methods were originally designed for two-class problems. We do not expect a sample to be classified into all classes with equal probability Such cases can hardly be differentiated by the generalized measures computed based on converted two-class problems. Besides taking into account both the classification accuracy and class discrimination power of classifiers, probabilistic confusion entropy tries to measure if samples are classified into true classes with high probabilities and into other classes.

Related Work
Performance Evaluation of Classification Models
Confusion Entropy Based on Probabilistic Confusion Matrix
Rules for Comparing Different Performance Measures
Conclusions
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.