Abstract

Continuous learning occurs naturally in human beings. However, Deep Learning methods suffer from a problem known as Catastrophic Forgetting (CF) that consists of a model drastically decreasing its performance on previously learned tasks when it is sequentially trained on new tasks. This situation, known as task interference, occurs when a network modifies relevant weight values as it learns a new task. In this work, we propose two main strategies to face the problem of task interference in convolutional neural networks. First, we use a sparse coding technique to adaptively allocate model capacity to different tasks avoiding interference between them. Specifically, we use a strategy based on group sparse regularization to specialize groups of parameters to learn each task. Afterward, by adding binary masks, we can freeze these groups of parameters, using the rest of the network to learn new tasks. Second, we use a meta learning technique to foster knowledge transfer among tasks, encouraging weight reusability instead of overwriting. Specifically, we use an optimization strategy based on episodic training to foster learning weights that are expected to be useful to solve future tasks. Together, these two strategies help us to avoid interference by preserving compatibility with previous and future weight values. Using this approach, we achieve state-of-the-art results on popular benchmarks used to test techniques to avoid CF. In particular, we conduct an ablation study to identify the contribution of each component of the proposed method, demonstrating its ability to avoid retroactive interference with previous tasks and to promote knowledge transfer to future tasks.

Highlights

  • A MONG the cognitive abilities of humans, memory is one of the most relevant

  • In the context of fewshot learning [16], [45], meta-learning techniques have been used to promote the learning of weights that can be quickly adapted to handle new tasks [10], [43], [48]. Taking inspiration from these ideas, we propose a meta-learning strategy that fosters learning of weights that can be useful to favor a positive transfer of knowledge between tasks, facilitating the acquisition of continual learning skills

  • We can observe that GoCaT improves the accuracy by reducing the gap in all tasks, while at the same time avoiding task interference

Read more

Summary

INTRODUCTION

A MONG the cognitive abilities of humans, memory is one of the most relevant. In effect, the ability to recall past experiences, knowledge, friendships, and emotions, is rooted in the essence of what makes us humans. We propose a new method that exploits two complementary learning strategies to mitigate CF in ANN, in particular, deep Convolutional Neural Network (CNN) These learning strategies are based on: i) dynamic allocation of model capacity to avoid interference between tasks, and ii) knowledge transfer from previous to new tasks to foster weight reusability instead of overwriting. A downside of this strategy is that the masks are learned independently of network parameters, leading to suboptimal solutions Another popular strategy is to use memory replay to recall critical information about previous datasets [21], [47], [48], either through saving elements of previous tasks or training GANs. to generate those elements past elements [17], [31], [52]. By fostering the recycling of parameters from previous tasks, we implement an efficient use of previous knowledge to support new tasks

META-LEARNING
METHOD DESCRIPTION
AVOIDING INTERFERENCE AMONG TASKS
EXPERIMENTAL EVALUATION
BASELINES
IMPLEMENTATION DETAILS
RESULTS
Findings
CONCLUSIONS
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call