论文标题
利用时间连贯性进行自我监督的一声视频重新识别
Exploiting Temporal Coherence for Self-Supervised One-shot Video Re-identification
论文作者
论文摘要
尽管重新识别的监督技术非常有效,但对大量注释的需求使它们对于大型相机网络不切实际。单发重新识别,该识别使用单个标记的曲目以及一个未标记的轨道池,是减少这种标签工作的潜在候选人。当前的一次性重新识别方法通过对标记和未标记数据之间的相互关系进行建模,但无法完全利用未标记数据本身池中存在的这种关系。在本文中,我们提出了一个名为“时间一致性”渐进式学习的新框架,该框架将时间连贯性用作一种新颖的自我监督辅助任务,以单次学习范式在未标记的轨道之间捕获这种关系。优化两个新的损失,这些损失在本地和全球规模上执行一致性,我们的框架可以学习更丰富,更具歧视性的表示。对两个具有挑战性的视频重新识别数据集进行了广泛的实验-MARS和DUKEMTMC-VEDEOREID-表明,我们所提出的方法能够更准确地估算未标记数据的真实标签,最高为8%\%$,并获得与现有的现有正式技术相比,获得明显更好的重新识别性能。
While supervised techniques in re-identification are extremely effective, the need for large amounts of annotations makes them impractical for large camera networks. One-shot re-identification, which uses a singular labeled tracklet for each identity along with a pool of unlabeled tracklets, is a potential candidate towards reducing this labeling effort. Current one-shot re-identification methods function by modeling the inter-relationships amongst the labeled and the unlabeled data, but fail to fully exploit such relationships that exist within the pool of unlabeled data itself. In this paper, we propose a new framework named Temporal Consistency Progressive Learning, which uses temporal coherence as a novel self-supervised auxiliary task in the one-shot learning paradigm to capture such relationships amongst the unlabeled tracklets. Optimizing two new losses, which enforce consistency on a local and global scale, our framework can learn learn richer and more discriminative representations. Extensive experiments on two challenging video re-identification datasets - MARS and DukeMTMC-VideoReID - demonstrate that our proposed method is able to estimate the true labels of the unlabeled data more accurately by up to $8\%$, and obtain significantly better re-identification performance compared to the existing state-of-the-art techniques.