Abstract

Classification is a fundamental problem in machine learning and data mining. During the past decades, numerous classification methods have been presented based on different principles. However, most existing classifiers cast the classification problem as an optimization problem and do not address the issue of statistical significance. In this paper, we formulate the binary classification problem as a two-sample testing problem. More precisely, our classification model is a generic framework that is composed of two steps. In the first step, the distance between the test instance and each training instance is calculated to derive two distance sets. In the second step, the two-sample test is performed under the null hypothesis that the two sets of distances are drawn from the same cumulative distribution. After these two steps, we have two ${p}$ -values for each test instance and the test instance is assigned to the class associated with the smaller ${p}$ -value. Essentially, the presented classification method can be regarded as an instance-based classifier based on hypothesis testing. The experimental results on 38 real data sets show that our method is able to achieve the same level performance as several classic classifiers and has significantly better performance than existing testing-based classifiers. Furthermore, we can handle outlying instances and control the false discovery rate of test instances assigned to each class under the same framework.

Highlights

  • Classification is a fundamental data analysis procedure, which is ubiquitously used across different fields

  • Based on the above observations, we present a new testingbased classification formulation, in which the null hypothesis is that, informally, the test instance does not belong to any class

  • The testing-based classification model has the advantage of controlling the false discovery rate (FDR) of classified test instances and handling outlying instances under the same framework

Read more

Summary

INTRODUCTION

Classification is a fundamental data analysis procedure, which is ubiquitously used across different fields. Based on the above observations, we present a new testingbased classification formulation, in which the null hypothesis is that, informally, the test instance does not belong to any class. It is very easy to control the type I error in terms of FDR in our formulation since the p-values of each test instance with respect to different classes will be generated in the classification phase. In other words, such testing-based classification formulation provides a unified framework to control the asymmetric classification error in a natural way. We can assign the test instance to the class that has the smallest p-value

K-NN VARIANTS
THE CHOICE OF TESTING METHODS
HANDLING OUTLIERS AND FDR CONTROL
RELATIONSHIP TO OTHER APPROACHES
CONNECTION TO NEAREST CENTROID CLASSIFIER
Findings
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call