论文标题

部分可观测时空混沌系统的无模型预测

A simple but strong baseline for online continual learning: Repeated Augmented Rehearsal

论文作者

Zhang, Yaqian, Pfahringer, Bernhard, Frank, Eibe, Bifet, Albert, Lim, Nick Jin Sean, Jia, Yunzhe

论文摘要

在线持续学习(OCL)旨在通过单个通过数据来逐步训练神经网络。基于彩排的方法试图用较小的内存近似观察到的输入分布,并以后重新审视它们以避免忘记。尽管具有强烈的经验表现,但排练方法仍然遭受过去数据损失景观和记忆样本的损失差异的差异。本文重新审视了在线设置中的排练动态。我们从偏见和动态的经验风险最小化的角度提供了关于内在内存过度符合风险的理论见解,并检查重复彩排的优点和限制。受我们的分析的启发,一个简单而直观的基线,重复的增强彩排(RAR)旨在解决在线彩排的拟合不足的困境。令人惊讶的是,在四个相当不同的OCL基准测试中,这种简单的基线表现优于香草排练9%-17%,并且显着改善了基于最新的彩排方法mir,Aser和Scr。我们还证明,RAR成功地实现了过去数据的损失格局和其学习轨迹中的高损失脊厌恶的准确近似。进行了广泛的消融研究,以研究重复和增强彩排和增强学习(RL)之间的相互作用(RL),以动态调整RAR的超参数,以平衡在线稳定性 - 塑性权衡权衡。代码可在https://github.com/yaqianzhang/repeatedeedaugmentedrearearsal上获得

Online continual learning (OCL) aims to train neural networks incrementally from a non-stationary data stream with a single pass through data. Rehearsal-based methods attempt to approximate the observed input distributions over time with a small memory and revisit them later to avoid forgetting. Despite its strong empirical performance, rehearsal methods still suffer from a poor approximation of the loss landscape of past data with memory samples. This paper revisits the rehearsal dynamics in online settings. We provide theoretical insights on the inherent memory overfitting risk from the viewpoint of biased and dynamic empirical risk minimization, and examine the merits and limits of repeated rehearsal. Inspired by our analysis, a simple and intuitive baseline, Repeated Augmented Rehearsal (RAR), is designed to address the underfitting-overfitting dilemma of online rehearsal. Surprisingly, across four rather different OCL benchmarks, this simple baseline outperforms vanilla rehearsal by 9%-17% and also significantly improves state-of-the-art rehearsal-based methods MIR, ASER, and SCR. We also demonstrate that RAR successfully achieves an accurate approximation of the loss landscape of past data and high-loss ridge aversion in its learning trajectory. Extensive ablation studies are conducted to study the interplay between repeated and augmented rehearsal and reinforcement learning (RL) is applied to dynamically adjust the hyperparameters of RAR to balance the stability-plasticity trade-off online. Code is available at https://github.com/YaqianZhang/RepeatedAugmentedRehearsal

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源