Abstract

Consider a domain-adaptive supervised learning setting, where a classifier learns from labeled data in a source domain and unlabeled data in a target domain to predict the corresponding target labels. If the classifier’s assumption on the relationship between domains (e.g. covariate shift, common subspace, etc.) is valid, then it will usually outperform a non-adaptive source classifier. If its assumption is invalid, it can perform substantially worse. Validating assumptions on domain relationships is not possible without target labels. We argue that, in order to make domain-adaptive classifiers more practical, it is necessary to focus on robustness; robust in the sense that an adaptive classifier will still perform at least as well as a non-adaptive classifier without having to rely on the validity of strong assumptions. With this objective in mind, we derive a conservative parameter estimation technique, which is transductive in the sense of Vapnik and Chervonenkis, and show for discriminant analysis that the new estimator is guaranteed to achieve a lower risk on the given target samples compared to the source classifier. Experiments on problems with geographical sampling bias indicate that our parameter estimator performs well.

Highlights

  • Generalization in supervised learning relies on the fact that future samples originate from the same underlying data-generating distribution as the ones used for training

  • Consider a domain-adaptive supervised learning setting, where a classifier learns from labeled data in a source domain and unlabeled data in a target domain to predict the corresponding target labels

  • Target Contrastive Pessimistic risk (TCP)-quadratic discriminant analysis (QDA) is better than S-QDA in eleven of the twelve

Read more

Summary

Introduction

Generalization in supervised learning relies on the fact that future samples originate from the same underlying data-generating distribution as the ones used for training. This is not the case in settings where data is collected from different locations, different measurement instruments are used or there is only access to biased data [25]. In these situations the labeled data does not represent the distribution of interest. We formulate a conservative adaptive classifier that always performs at least as well as the non-adaptive one.

Objectives
Methods
Results
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call