Abstract

Catastrophic forgetting is a well-known characteristic of diverse parameterized supervised learning models. Artificial Neural Networks (ANN) face severe catastrophic forgetting or catastrophic interference in the sequential learning of information that is intolerable to both engineering applications and the human memory model. This problem can be solved through the machine learning community. The main intent of this paper is to implement a novel continuous learning model that could overcome the problem of catastrophic forgetting. The stream data from diverse benchmark sources are used for experimenting. The proposed model adopts the Deep Neural Network (DNN) with testing weight updates. When new data are given, the hybrid meta-heuristic algorithm helps to update the weight that manages to prevent the learning from catastrophic forgetting. The weight update strategy for solving the catastrophic forgetting is done by the hybridization of Shark Smell Optimization (SSO) and Jaya Algorithm (JA) that is named as Hybrid Shark Smell with Jaya Optimization (HSS-JO). A multi-objective function concerning the parameters associated with catastrophic forgetting like accuracy and remembering is used for framing the proposed continuous learning. The proof of the proposed model over conventional models on three publicly available datasets is given for final validation. From the analysis, in Table 3, the accuracy of the proposed HSS-JO-DNN is 38%, 0.5%, 0.55%, 0.14%, 0.98%, and 0.5% superior to DT, NB, KNN, SVM, NN, and RNN, respectively. For dataset 2, the FNR of the proposed HSS-JO-DNN is 90% superior to DT, 89% improved than NB, KNN, and NN, 94% improved than SVM, and 24% improved than RNN. Similarly, the proposed HSS-JO performance is higher in all the results. Thus, the proposed HSS-JO performs better than all the other traditional methods.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call