论文标题
基于注意力的跨层域对准无监督域的适应
Attention-based Cross-Layer Domain Alignment for Unsupervised Domain Adaptation
论文作者
论文摘要
无监督的域适应性(UDA)旨在从标记的源域学习可转移的知识,并将训练有素的模型调整为未标记的目标域。为了弥合源域和目标域之间的差距,一种流行的策略是通过对齐深层模型提取的语义特征来最大程度地减少分布差异。现有基于对齐的方法主要集中于减少同一模型层中的域差异。但是,由于域移动,相同级别的语义信息可以在模型层上分布。为了进一步提高模型适应性性能,我们提出了一种新的方法,称为“基于注意力的跨层域对齐(ACDA)”,该方法通过模型层捕获了源和目标域之间的语义关系,并通过动态注意机制自动校准每个语义信息。精心设计的注意机制旨在根据其语义相似性以精确的域对齐方式重新加权每个跨层对,从而有效地匹配模型适应过程中的每个语义信息。在多个基准数据集上进行的广泛实验始终表明,所提出的ACDA会产生最先进的性能。
Unsupervised domain adaptation (UDA) aims to learn transferable knowledge from a labeled source domain and adapts a trained model to an unlabeled target domain. To bridge the gap between source and target domains, one prevailing strategy is to minimize the distribution discrepancy by aligning their semantic features extracted by deep models. The existing alignment-based methods mainly focus on reducing domain divergence in the same model layer. However, the same level of semantic information could distribute across model layers due to the domain shifts. To further boost model adaptation performance, we propose a novel method called Attention-based Cross-layer Domain Alignment (ACDA), which captures the semantic relationship between the source and target domains across model layers and calibrates each level of semantic information automatically through a dynamic attention mechanism. An elaborate attention mechanism is designed to reweight each cross-layer pair based on their semantic similarity for precise domain alignment, effectively matching each level of semantic information during model adaptation. Extensive experiments on multiple benchmark datasets consistently show that the proposed method ACDA yields state-of-the-art performance.