Abstract

In order to overcome many inherent defects of support vector machine (SVM), for example, the kernel function must satisfy the Mercer condition, relevance vector machine (RVM) was proposed to avoid these shortcomings of SVM This study concerns with the performance of RVM and SVM for regression and classification problem. Because RVM is based on Bayesian framework, a priori knowledge of the penalty term is introduced, the RVM needless relevance vectors (RVs) (support vectors (SVs) in SVM) but better generalization ability than SVM. In this paper, Sparse Bayesian learning (SBL) is firstly introduced and then RVM regression and classification models which based on SBL are introduced secondly, and then by inference the parameters, the RVM learning is transform into maximize the marginal likelihood function estimation, and give three kinds of commonly used estimation methods. Finally, we do some simulation experiments to show that the RVM has less RVs or SVs but better generalization ability than SVM whether regression or classification case, and also show that different kernel functions will impact the performance of RVM. However, there does not exist the performance of a kernel function is much better than other kernel functions.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call