Abstract

Primitive machine learning method such as Support Vector Machine (SVM) or k-Nearest Neighbor (k-NN) faces a major challenge when its training and test data is distributed with large-scale variations in lighting conditions, color, backgrounds, size, etc. The variation may be because training and testing data can come from related but some other domains. Considerable efforts have been made in the development of transfer learning methods. However, most current work focuses only on the following goals or objectives: i)Preservation of source domain discriminative information with Linear Discriminant Analysis (LDA); ii)Maximization of target domain variance; iii) Subspace alignment; iv) Minimization of marginal and conditional distribution by using the Maximum Mean Discrepancy (MMD) criterion;v)Preservation of original similarity of the data samples. Current approaches to preserve source domain discriminant information can easily misclassify the target domain samples which are distributed near the edge of the cluster. In order to overcome the limitations of existing transfer learning methods, we propose a novel Unsupervised Transfer Learning Via Relative Distance Comparisons (UTRDC) method. UTRDC optimizes all the aforementioned objectives jointly with a common projection vector matrix for both domains as well as uses the relative distance constraints for better inter-class separability and intra-class compactness. Furthermore, we extend our proposed method UTRDC to a kernelized version to deal with non-linear separable datasets. Extensive experimentation on two real-world problems datasets (PIE face and Office+Caltech) has proven that the proposed methods outperform several approaches to non-transfer learning and transfer learning.

Highlights

  • Supervised learning methods (such as Support Vector Machine (SVM), k-Nearest Neighbor (k-NN), etc.) [1]–[4] are arguably the most common types of machine learning and have been very successful in various real-world applications

  • We introduce a Transfer Learning (TL) method called Unsupervised Transfer learning via Relative Distance Comparisons (UTRDC) for reducing the distribution difference between both domains

  • In order to deal with the non-linear datasets, we extend our proposed method to a kernelized method called Kernel Unsupervised Transfer learning via Relative Distance Comparisons (KUTRDC)

Read more

Summary

INTRODUCTION

Supervised learning methods (such as Support Vector Machine (SVM), k-Nearest Neighbor (k-NN), etc.) [1]–[4] are arguably the most common types of machine learning and have been very successful in various real-world applications. Previous studies about TL focus on satisfying one or more of the following objectives: i)Preserving discriminative information; ii)Subspace alignment; iii)Preserving marginal and conditional distributions using Maximum Mean Discrepancy (MMD) criterion; iv)Maximizing target domain variance;v) preserving original similarity of data samples using Laplacian term, to minimize the distribution divergence or difference between both domains. We don’t have label information for the target domain for satisfying those conditions Remember that both domains are highly correlated and have the same distribution in the common feature space. We introduce a TL method called Unsupervised Transfer learning via Relative Distance Comparisons (UTRDC) for reducing the distribution difference between both domains. Our proposed UTRDC algorithm finds a common projection vector for both domains while satisfying the following objectives: i)Preserving discriminative information by learning a relative distance metric; ii)Preserving marginal and conditional distributions; iii) maximizing target domain variance; iv) preserving original similarity of data samples. They achieved 88.97% and 88.32% accuracy, respectively, for the Office+Caltech using VGG-Net features data-set

RELATED WORK
NOTATIONS AND PROBLEM DEFINITION
UNSUPERVISED TRANSFER LEARNING VIA RELATIVE DISTANCE COMPARISONS
FORMULATION OF PROPOSED OVERALL OBJECTIVE FUNCTION
KERNELIZATION OF PROPOSED OVERALL OBJECTIVE
Construct marginal and conditional distribution
EXPERIMENTS
COMPLEXITY ANALYSIS OF OUR PROPOSED METHODS
Findings
CONCLUSION AND FUTURE WORK
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call