论文标题
重新访问机器人学习中的对抗性鲁棒性 - 准确性的权衡
Revisiting the Adversarial Robustness-Accuracy Tradeoff in Robot Learning
论文作者
论文摘要
对抗性训练(即,对对抗性扰动输入数据进行培训)是一种良好的方法,可在推理过程中使神经网络稳健地对潜在的对抗性攻击。但是,改善的鲁棒性不是免费的,而是伴随着整体模型准确性和性能的降低。最近的工作表明,在实用的机器人学习应用中,对抗训练的影响不会构成公平的权衡,而是在整体机器人性能中衡量时会造成净损失。这项工作通过系统地分析强大的训练方法和理论的最新进展以及与对抗性机器人学习的结合,可以使机器人学习中的稳健性 - 准确性权衡取舍,能够使对抗性培训适合于现实世界中的机器人应用。我们评估了三个不同的机器人学习任务,从适合使用的高保真环境中的自动驾驶到移动机器人导航和手势识别。我们的结果表明,尽管这些技术以相对规模的折衷进行了逐步改进,但对对抗性训练引起的名义准确性的负面影响仍然超过了稳健性的改善,从而提高了数量级。我们得出的结论是,尽管进步正在发生,但在实践中有益于机器人学习任务之前,必须进一步进步。
Adversarial training (i.e., training on adversarially perturbed input data) is a well-studied method for making neural networks robust to potential adversarial attacks during inference. However, the improved robustness does not come for free but rather is accompanied by a decrease in overall model accuracy and performance. Recent work has shown that, in practical robot learning applications, the effects of adversarial training do not pose a fair trade-off but inflict a net loss when measured in holistic robot performance. This work revisits the robustness-accuracy trade-off in robot learning by systematically analyzing if recent advances in robust training methods and theory in conjunction with adversarial robot learning, are capable of making adversarial training suitable for real-world robot applications. We evaluate three different robot learning tasks ranging from autonomous driving in a high-fidelity environment amenable to sim-to-real deployment to mobile robot navigation and gesture recognition. Our results demonstrate that, while these techniques make incremental improvements on the trade-off on a relative scale, the negative impact on the nominal accuracy caused by adversarial training still outweighs the improved robustness by an order of magnitude. We conclude that although progress is happening, further advances in robust learning methods are necessary before they can benefit robot learning tasks in practice.