论文标题
SS-IL:分开的软智能学习
SS-IL: Separated Softmax for Incremental Learning
论文作者
论文摘要
我们考虑了类增量学习(CIL)问题,其中学习代理人通过逐步到达培训数据批次不断学习新课程,并旨在在到目前为止所学的所有课程中很好地预测。问题的主要挑战是灾难性的遗忘,对于基于示例性记忆的CIL方法,人们普遍知道,由于新类和旧类之间的数据失衡,遗忘通常是由分类评分偏差引起的(在示例性模拟中)。尽管已经提出了几种方法来通过一些其他后处理(例如,得分重新缩放或平衡的微调)来纠正这种分数偏见,但没有就此类偏见的根本原因进行系统分析。为此,我们分析了通过组合所有旧类和新类的输出得分来计算软磁概率的主要原因。然后,我们提出了一种新方法,称为分离的软智能学习(SS-IL),该方法由分离的SoftMax(SS)输出层组成,结合了任务知识蒸馏(TKD)来解决此类偏见。在几个大规模CIL基准数据集的广泛实验结果中,我们通过在没有任何额外的后处理的情况下达到更加平衡的预测分数来表明我们的SS-IL实现了强大的最先进的准确性。
We consider class incremental learning (CIL) problem, in which a learning agent continuously learns new classes from incrementally arriving training data batches and aims to predict well on all the classes learned so far. The main challenge of the problem is the catastrophic forgetting, and for the exemplar-memory based CIL methods, it is generally known that the forgetting is commonly caused by the classification score bias that is injected due to the data imbalance between the new classes and the old classes (in the exemplar-memory). While several methods have been proposed to correct such score bias by some additional post-processing, e.g., score re-scaling or balanced fine-tuning, no systematic analysis on the root cause of such bias has been done. To that end, we analyze that computing the softmax probabilities by combining the output scores for all old and new classes could be the main cause of the bias. Then, we propose a new method, dubbed as Separated Softmax for Incremental Learning (SS-IL), that consists of separated softmax (SS) output layer combined with task-wise knowledge distillation (TKD) to resolve such bias. Throughout our extensive experimental results on several large-scale CIL benchmark datasets, we show our SS-IL achieves strong state-of-the-art accuracy through attaining much more balanced prediction scores across old and new classes, without any additional post-processing.