Abstract
K-SVCR and Twin-KSVC are two novel algorithms to deal with multi-class problems. They have achieved good performance since they evaluate all training samples into a “1-versus-1-versus-rest” structure. But they are extremely time consuming, so it remains challenging to apply them into large-scale problems directly. Inspired by the sparse solution of SVMs, in this paper, we propose a safe sample elimination rule (SSE) for multi-class classifiers K-SVCR and Twin-KSVC, termed as SSE-K-SVCR and SSE-T-KSVC, to reduce computation time. With our rule, many redundant samples of all classes can be identified and deleted before actually solving the problem, so the scale of dual problems can be reduced a lot. And our methods are safe, i.e., they can derive identical optimal solutions as K-SVCR and Twin-KSVC, respectively. So the testing accuracy keeps unchanged. Besides, the methods can be embedded into grid search method to accelerate the whole training process, and they are effective both for penalty parameter and kernel parameter. Finally, a fast algorithm clipDCD is introduced to reduce the computation time for larger datatset. Experimental results on one artificial dataset and seventeen benchmark datasets demonstrate the effectiveness and safety of our proposed methods.
Published Version
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.