论文标题
几乎没有学习作为域适应的学习:算法和分析
Few-Shot Learning as Domain Adaptation: Algorithm and Analysis
论文作者
论文摘要
为了识别只有很少样本的看不见的课程,很少的学习(FSL)使用从所见类中学到的先验知识。 FSL的一个主要挑战是,看不见的类别的分布与所见类别的分布不同,即使在可见类中进行了元训练的模型,也会导致泛化。这种类差异引起的分布变化可以被视为域移动的特殊情况。在本文中,我们首次提出了一个具有注意力(DAPNA)的域适应性典型网络,以在元学习框架中明确解决此类域移位问题。具体而言,我们由基于设定的变压器的注意模块武装,我们用两个没有类别重叠的子剧本构建了每个情节,以模拟可见和看不见的类之间的域移动。为了使两个子剧集的特征分布与有限的训练样本保持一致,使用特征转移网络以及边距差异差异(MDD)损失。重要的是,提供了理论分析以赋予我们DAPNA的学习结构。广泛的实验表明,我们的DAPNA通常优于最先进的FSL替代方案,通常是通过大量的边缘。
To recognize the unseen classes with only few samples, few-shot learning (FSL) uses prior knowledge learned from the seen classes. A major challenge for FSL is that the distribution of the unseen classes is different from that of those seen, resulting in poor generalization even when a model is meta-trained on the seen classes. This class-difference-caused distribution shift can be considered as a special case of domain shift. In this paper, for the first time, we propose a domain adaptation prototypical network with attention (DAPNA) to explicitly tackle such a domain shift problem in a meta-learning framework. Specifically, armed with a set transformer based attention module, we construct each episode with two sub-episodes without class overlap on the seen classes to simulate the domain shift between the seen and unseen classes. To align the feature distributions of the two sub-episodes with limited training samples, a feature transfer network is employed together with a margin disparity discrepancy (MDD) loss. Importantly, theoretical analysis is provided to give the learning bound of our DAPNA. Extensive experiments show that our DAPNA outperforms the state-of-the-art FSL alternatives, often by significant margins.