Abstract

Two major problems are encountered when using fuzzy SVM: (a) the number of local minima increases exponentially with the number of samples and (b) the quantity of required computer storage, required for a regular quadratic programming solver, increases by an exponential magnitude as the problem size expands. The Kernel-Adatron family of algorithms gaining attention lately which has allowed to handle very large classification and regression problems. However, these methods treat different types of samples (Noise, border, and core) with the same manner, which causes searches in unpromising areas and increases the number of iterations. In this work, we introduce a hybrid method to overcome these shortcoming, namely Optimal Recurrent Neural Network Density Based fuzzy Support Vector Machine (Opt-RNN-DBFSVM). This method consists of four steps: (a) characterization of different samples, (b) elimination of samples with a low probability of being a support vector, (c) construction of an appropriate recurrent neural network based on an original energy function, and (d) solution of the system of differential equations, managing the dynamics of the RNN, using the Euler–Cauchy method involving an optimal time step. Thanks to its recurrent architecture, the RNN remembers the regions explored during the search process. We demonstrated that RNN-FSVM converges to feasible support vectors and Opt-RNN-DBFSVM has a very low time complexity compared to RNN-FSVM with constant time step, and KAs-FSVM. Several experiments were performed on academic data sets. We used several classification performance measures to compare Opt-RNN-DBFSVM to different classification methods and the results obtained show the good performance of the proposed method.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call