论文标题
持续学习反对对抗攻击的敏感性
Susceptibility of Continual Learning Against Adversarial Attacks
论文作者
论文摘要
最近的持续学习方法主要集中于减轻灾难性遗忘。然而,两个关键领域仍然相对尚未探索:1)评估提议的方法的鲁棒性和2)确保学习任务的安全性。本文调查了不断学习的任务的敏感性,包括当前和先前获得的任务对对抗性攻击的敏感性。具体而言,我们已经观察到,属于任何任务的任何类都可以轻松定位和错误分类为任何其他任务的所需目标类。这种敏感性或对对抗性攻击的任务的脆弱性引起了人们对数据完整性和隐私的深刻关注。为了评估持续学习方法的鲁棒性,我们考虑了在所有三种情况下的持续学习方法,即任务收入学习,领域的收入学习和班级学习。在这方面,我们探讨了三种基于正则化的方法,三种基于重播的方法的鲁棒性以及一种结合重播和示例方法的混合技术。我们从经验上证明,在任何持续学习的环境中,任何阶级,无论是属于当前的任务还是以前学习的任务,都容易被错误分类。我们的观察结果确定了针对对抗性攻击的持续学习方法的潜在局限性,并强调当前的持续学习算法不适合在现实世界中部署。
Recent continual learning approaches have primarily focused on mitigating catastrophic forgetting. Nevertheless, two critical areas have remained relatively unexplored: 1) evaluating the robustness of proposed methods and 2) ensuring the security of learned tasks. This paper investigates the susceptibility of continually learned tasks, including current and previously acquired tasks, to adversarial attacks. Specifically, we have observed that any class belonging to any task can be easily targeted and misclassified as the desired target class of any other task. Such susceptibility or vulnerability of learned tasks to adversarial attacks raises profound concerns regarding data integrity and privacy. To assess the robustness of continual learning approaches, we consider continual learning approaches in all three scenarios, i.e., task-incremental learning, domain-incremental learning, and class-incremental learning. In this regard, we explore the robustness of three regularization-based methods, three replay-based approaches, and one hybrid technique that combines replay and exemplar approaches. We empirically demonstrated that in any setting of continual learning, any class, whether belonging to the current or previously learned tasks, is susceptible to misclassification. Our observations identify potential limitations of continual learning approaches against adversarial attacks and highlight that current continual learning algorithms could not be suitable for deployment in real-world settings.