Abstract

Kernel logistic regression is a widely used method in machine learning applications. However, the common kernels are not flexible or representative enough due to their simplistic parametric formulation. Non-parametric learning on kernels can enlarge the representation capability of the kernel, but the low-rank property is required to suppress the model complexity. To improve the flexibility of kernels, we perform low-rank decomposition on an adjust matrix, which can explicitly control the model complexity without using complicated rank penalty terms. In this paper, we also extend our method to learn an indefinite kernel by a different decomposition. Experimental results demonstrate that the proposed learning method for non-parametric kernels outperforms other representative algorithms.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call