Abstract

AbstractWhile the original SVM seeks the discriminative plane that maximizes the margin in the feature space (the Hilbert space), this paper investigates the framework that maximizes the margin in the input space. This framework is considered to be effective for cases in which a priori knowledge is embedded as input space estimates. In the approach taken in this paper, approximating the margin in the input space by Taylor expansion is essential. The algorithm obtained is a kind of alternating optimization comprising the step of obtaining projections onto the discriminative plane from sample points by Newton's method and the step of determining parameters of the discriminative plane by convex quadratic programming. The algorithm converges to a stably local optimal solution under comparatively lenient conditions. In addition, the optimization problem to be solved includes the original SVM as a special case. However, since the amount of computation increases as the dimensions of the input space increase in the proposed algorithm, this paper proposes a simplified algorithm obtained by combining the original SVM and the abridged proposed algorithm. © 2004 Wiley Periodicals, Inc. Syst Comp Jpn, 35(14): 78–86, 2004; Published online in Wiley InterScience (www.interscience.wiley.com). DOI 10.1002/scj.10631

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.