论文标题
学习图像标签,用于培训强大的分类模型
Learning Image Labels On-the-fly for Training Robust Classification Models
论文作者
论文摘要
当前的深度学习范例在很大程度上受益于大量注释数据。但是,注释的质量通常在标签中有所不同。已经进行了多观察者研究来研究这些注释差异(通过将相同的数据标记多次标记)及其对医学图像分析(例如医学图像分析)的影响。这个过程确实为已经乏味的注释工作增加了额外的负担,通常需要在特定领域中进行专业培训和专业知识。另一方面,基于NLP算法的自动注释方法最近显示了有望是合理的替代方案,依赖于临床系统中广泛使用的这些图像的现有诊断报告。与人类标记相比,不同的算法为标签提供了不同的品质,甚至更嘈杂。在本文中,我们展示了如何一起利用嘈杂的注释(例如,来自不同算法的标签)并相互利用分类任务的学习。具体而言,引入了注意力标签的概念,以在训练数据时在fly上样本进行样品。基于元训练的标签采样模块旨在参加通过其他背部传播过程中最大程度地学习模型学习的标签。我们将注意力贴在标签方案上应用于合成噪声CIFAR-10数据集的分类任务来证明该概念,然后在胸部X射线图像上展示了来自医院规模的数据集(MIMIC-CXR)(MIMIC-CXR)和手持式数据集合(openabeled pardigig)的胸部X射线图像上的较高结果(在多种疾病分类AUC中平均增加了3-5%)。
Current deep learning paradigms largely benefit from the tremendous amount of annotated data. However, the quality of the annotations often varies among labelers. Multi-observer studies have been conducted to study these annotation variances (by labeling the same data for multiple times) and its effects on critical applications like medical image analysis. This process indeed adds an extra burden to the already tedious annotation work that usually requires professional training and expertise in the specific domains. On the other hand, automated annotation methods based on NLP algorithms have recently shown promise as a reasonable alternative, relying on the existing diagnostic reports of those images that are widely available in the clinical system. Compared to human labelers, different algorithms provide labels with varying qualities that are even noisier. In this paper, we show how noisy annotations (e.g., from different algorithm-based labelers) can be utilized together and mutually benefit the learning of classification tasks. Specifically, the concept of attention-on-label is introduced to sample better label sets on-the-fly as the training data. A meta-training based label-sampling module is designed to attend the labels that benefit the model learning the most through additional back-propagation processes. We apply the attention-on-label scheme on the classification task of a synthetic noisy CIFAR-10 dataset to prove the concept, and then demonstrate superior results (3-5% increase on average in multiple disease classification AUCs) on the chest x-ray images from a hospital-scale dataset (MIMIC-CXR) and hand-labeled dataset (OpenI) in comparison to regular training paradigms.