Abstract

Multiple kernel learning (MKL) serves as an attractive research direction in current kernel machine learning field. It can flexibly process diverse characteristics of patterns such as heterogeneous information or irregular data, non-flat distribution of high-dimensional samples etc. The existing MKL models are usually built on SVM. However, there is still potential to improve the performance of MKL instead of learning based on SVM. Nonparallel support vector machine (NPSVM), as a novel classifier, pursues two nonparallel proximal hyperplanes with several incomparable advantages over the state-of-the-art classifiers. In this paper, we propose a new model termed as MKNPSVM for classification. By integrating NPSVM into the MKL framework, MKNPSVM inherits the advantages of them and opens a new perspective to extend NPSVM to the MKL field. To solve MKNPSVM efficiently, we provide an alternating optimization algorithm (Alter-MKNPSVM for short) as the solution. We theoretically analyze the performance of MKNPSVM from three viewpoints: the generalization capability analysis, the convergence analysis and the comparisons with NPSVM and MKL. Experimental results on eighteen publicly available UCI data sets confirm the effectiveness of our method.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call