Abstract
Kernel logistic regression is a widely used method in machine learning applications. However, the common kernels are not flexible or representative enough due to their simplistic parametric formulation. Non-parametric learning on kernels can enlarge the representation capability of the kernel, but the low-rank property is required to suppress the model complexity. To improve the flexibility of kernels, we perform low-rank decomposition on an adjust matrix, which can explicitly control the model complexity without using complicated rank penalty terms. In this paper, we also extend our method to learn an indefinite kernel by a different decomposition. Experimental results demonstrate that the proposed learning method for non-parametric kernels outperforms other representative algorithms.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.