论文标题
辅助知识对持续学习的影响
Effects of Auxiliary Knowledge on Continual Learning
论文作者
论文摘要
在持续学习(CL)中,对神经网络进行了培训,该数据流的分布会随着时间而变化。在这种情况下,主要问题是如何在不忘记旧知识(即灾难性遗忘)的情况下学习新信息。大多数现有的CL方法都集中在寻找保留获得知识的解决方案,因此要在模型的过去工作。但是,我们认为,由于模型必须不断学习新任务,因此专注于可以改善以下任务学习的当前知识也很重要。在本文中,我们提出了一种新的,简单的CL算法,该算法着重于解决当前任务的方式,以促进下一个任务。更具体地说,我们的方法将主要数据流与次要,多样和不相关的流相结合,网络可以从中吸引辅助知识。从不同角度来看,这有助于模型,因为辅助数据可能包含当前任务的有用功能,并且可以将传入的任务类映射到辅助类中。此外,在当前任务中添加数据是隐含的,因为我们正在迫使提取更具歧视性的特征,使分类器更加健壮。我们的方法可以在最常见的CL图像分类基准上胜过现有的最新模型。
In Continual Learning (CL), a neural network is trained on a stream of data whose distribution changes over time. In this context, the main problem is how to learn new information without forgetting old knowledge (i.e., Catastrophic Forgetting). Most existing CL approaches focus on finding solutions to preserve acquired knowledge, so working on the past of the model. However, we argue that as the model has to continually learn new tasks, it is also important to put focus on the present knowledge that could improve following tasks learning. In this paper we propose a new, simple, CL algorithm that focuses on solving the current task in a way that might facilitate the learning of the next ones. More specifically, our approach combines the main data stream with a secondary, diverse and uncorrelated stream, from which the network can draw auxiliary knowledge. This helps the model from different perspectives, since auxiliary data may contain useful features for the current and the next tasks and incoming task classes can be mapped onto auxiliary classes. Furthermore, the addition of data to the current task is implicitly making the classifier more robust as we are forcing the extraction of more discriminative features. Our method can outperform existing state-of-the-art models on the most common CL Image Classification benchmarks.