Abstract

In the present era of big data applications, incremental learning has emerged as the most admired area of research where ever-ending tasks from different application domains arrive temporarily, and the learning models majorly focus on newly arriving tasks while forgetting the historically acquired knowledge of old tasks. In this context, when a deep neural network (DNN) performs a sequence of learning tasks, it usually forgets the previously learned tasks. This phenomenon is known as catastrophic forgetting (CF), and the existing state-of-the-art deep learning approaches are unable to handle this issue. Following this, many DNN-based state-of-the-art solutions have been proposed to alleviate the problem that ensures a more natural way of human brain imitation. During training, deep learning algorithms are subjected to a random sequence of diversified datasets (also referred to as cross-domain datasets) that typically belong to multiple domains and have a variety of feature representation, distribution, density, etc. In such cases, CF may become a serious issue because the model tends to forget the previously learned features that affect the overall performance. Because there has been very little research in this area, a more robust incremental deep learning algorithm is proposed in our study to learn desired tasks incrementally. To overcome this situation, proper sequencing of cross-domain datasets is important so that related datasets are kept together to enhance the learnability of the model. Our research proposes a cross-domain order approach with hard attention to the task (HAT) methodology that uses an attention-gated mechanism to efficiently learn the task-specific features. Intense experiments have been performed, and the performance of the proposed approach is compared with the changing cross-domain orders with the baseline models over the following datasets: MNIST, P-MNIST, CIFAR-10, CIFAR-100, and SVHN, etc. The results of the experiments show that learning through cross-domain orders results in better incremental learning.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call