论文标题

通过数据镜头解释深层模型

Interpreting Deep Models through the Lens of Data

论文作者

Mercier, Dominique, Siddiqui, Shoaib Ahmed, Dengel, Andreas, Ahmed, Sheraz

论文摘要

识别与分类器相关的输入数据点(即用作支持向量)最近促使研究人员对可解释性以及数据集调试的兴趣。本文对方法进行了深入的分析,该方法试图识别这些数据点对所得分类器的影响。为了量化影响的质量,我们策划了一组实验,我们根据从不同方法获得的影响信息来调试和修剪数据集。为此,我们为分类器提供了标记错误的示例,这些示例阻碍了整体性能。因此,由于分类器是数据和模型的组合,因此,对于深度学习模型的可解释性,必须分析这些影响。对结果的分析表明,某些可解释性方法比使用随机方法更好地检测错误标签,但是与这些方法的主张相反,基于训练损失的样本选择表明表现出色。

Identification of input data points relevant for the classifier (i.e. serve as the support vector) has recently spurred the interest of researchers for both interpretability as well as dataset debugging. This paper presents an in-depth analysis of the methods which attempt to identify the influence of these data points on the resulting classifier. To quantify the quality of the influence, we curated a set of experiments where we debugged and pruned the dataset based on the influence information obtained from different methods. To do so, we provided the classifier with mislabeled examples that hampered the overall performance. Since the classifier is a combination of both the data and the model, therefore, it is essential to also analyze these influences for the interpretability of deep learning models. Analysis of the results shows that some interpretability methods can detect mislabels better than using a random approach, however, contrary to the claim of these methods, the sample selection based on the training loss showed a superior performance.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源