论文标题
DMT:半监督学习的动态相互培训
DMT: Dynamic Mutual Training for Semi-Supervised Learning
论文作者
论文摘要
最近的半监督学习方法使用伪监督作为核心思想,尤其是生成伪标签的自我训练方法。但是,伪标签是不可靠的。自我训练方法通常依赖于单个模型预测置信度来滤除较低的伪伪标签,从而保持较高的信心误差并浪费许多低信心正确的标签。在本文中,我们指出,模型很难反对自己的错误。取而代之的是,在不同模型之间利用模型间分歧是定位伪标签错误的关键。从这个新的角度来看,我们通过动态加权损耗函数(称为动态相互训练(DMT))提出了两个不同模型之间的相互训练。我们通过将两个不同模型的预测与训练中动态重量损失进行比较来量化模型间的分歧,在训练中,较大的分歧表明可能的误差,并且对应于较低的损失值。广泛的实验表明,DMT在图像分类和语义分段中都达到了最新的性能。我们的代码在https://github.com/voldemortx/dst-cbc上发布。
Recent semi-supervised learning methods use pseudo supervision as core idea, especially self-training methods that generate pseudo labels. However, pseudo labels are unreliable. Self-training methods usually rely on single model prediction confidence to filter low-confidence pseudo labels, thus remaining high-confidence errors and wasting many low-confidence correct labels. In this paper, we point out it is difficult for a model to counter its own errors. Instead, leveraging inter-model disagreement between different models is a key to locate pseudo label errors. With this new viewpoint, we propose mutual training between two different models by a dynamically re-weighted loss function, called Dynamic Mutual Training (DMT). We quantify inter-model disagreement by comparing predictions from two different models to dynamically re-weight loss in training, where a larger disagreement indicates a possible error and corresponds to a lower loss value. Extensive experiments show that DMT achieves state-of-the-art performance in both image classification and semantic segmentation. Our codes are released at https://github.com/voldemortX/DST-CBC .