Abstract

For semi-supervised techniques to be applied safely in practice we at least want methods to outperform their supervised counterparts. We study this question for classification using the well-known quadratic surrogate loss function. Unlike other approaches to semi-supervised learning, the procedure proposed in this work does not rely on assumptions that are not intrinsic to the classifier at hand. Using a projection of the supervised estimate onto a set of constraints imposed by the unlabeled data, we find we can safely improve over the supervised solution in terms of this quadratic loss. More specifically, we prove that, measured on the labeled and unlabeled training data, this semi-supervised procedure never gives a lower quadratic loss than the supervised alternative. To our knowledge this is the first approach that offers such strong, albeit conservative, guarantees for improvement over the supervised solution. The characteristics of our approach are explicated using benchmark datasets to further understand the similarities and differences between the quadratic loss criterion used in the theoretical results and the classification accuracy typically considered in practice.

Highlights

  • We consider the problem of semi-supervised classification using the quadratic loss function, which is known as least squares classification or Fisher’s linear discriminant classification (Hastie et al 2009; Poggio and Smale 2003)

  • We introduce a constraint set of parameter vectors induced by the unlabeled data, which does not rely on additional assumptions about the data

  • Our experiments indicate the results hold when performance is evaluated on objects in a test set that were not used as unlabeled objects during training

Read more

Summary

Introduction

We consider the problem of semi-supervised classification using the quadratic loss function, which is known as least squares classification or Fisher’s linear discriminant classification (Hastie et al 2009; Poggio and Smale 2003). Much work has been done on semi-supervised classification, in particular on what additional assumptions about the unlabeled data may help improve classification performance These additional assumptions, while successful in some settings, are less successful in others where they do not hold. We introduce a conservative approach to training a semi-supervised version of the least squares classifier that is guaranteed to improve over the supervised least squares classifier, in terms of the quadratic loss on the labeled and unlabeled examples. It is the first procedure for which it is possible to give strong guarantees of non-degradation of this type (Theorem 1)

Objectives
Findings
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call