论文标题
半监督的深度视角立体声
Semi-supervised Deep Multi-view Stereo
论文作者
论文摘要
在受监督和无监督的设置中,基于学习的多视图立体声(MV)中已经看到了重大进展。为了结合其在准确性和完整性方面的优点,同时减少了对昂贵标签数据的需求,本文在半监督的环境中探索了基于学习的MV的问题,该问题只有MVS数据的很小一部分与密集的深度地面真相相连。但是,由于场景和柔性设置的巨大变化,它可能会破坏经典半监督学习中的基本假设,即未标记的数据和标记的数据共享相同的标签空间和数据分布,称为半监督的MVS问题中的半监督分布差距歧义。为了解决这些问题,我们提出了一种新型的半监督分布式MVS框架,即SDA-MVS。对于基本假设在MVS数据中起作用的简单情况,一致性正则化鼓励模型预测在原始样本和随机增强样本之间保持一致。对于MVS数据中基本假设存在冲突的进一步麻烦的情况,我们提出了一种新型的样式一致性损失,以减轻分布差距引起的负面影响。未标记的样品的视觉样式被转移到标记的样品中以缩小差距,并且在原始标记的样品中使用标签进一步监督了生成样品的模型预测。多个MVS数据集的半监督设置的实验结果表明,该方法的出色性能。在骨干网络中使用相同的设置,我们提出的SDA-MV优于其完全监管和无监督的基线。
Significant progress has been witnessed in learning-based Multi-view Stereo (MVS) under supervised and unsupervised settings. To combine their respective merits in accuracy and completeness, meantime reducing the demand for expensive labeled data, this paper explores the problem of learning-based MVS in a semi-supervised setting that only a tiny part of the MVS data is attached with dense depth ground truth. However, due to huge variation of scenarios and flexible settings in views, it may break the basic assumption in classic semi-supervised learning, that unlabeled data and labeled data share the same label space and data distribution, named as semi-supervised distribution-gap ambiguity in the MVS problem. To handle these issues, we propose a novel semi-supervised distribution-augmented MVS framework, namely SDA-MVS. For the simple case that the basic assumption works in MVS data, consistency regularization encourages the model predictions to be consistent between original sample and randomly augmented sample. For further troublesome case that the basic assumption is conflicted in MVS data, we propose a novel style consistency loss to alleviate the negative effect caused by the distribution gap. The visual style of unlabeled sample is transferred to labeled sample to shrink the gap, and the model prediction of generated sample is further supervised with the label in original labeled sample. The experimental results in semi-supervised settings of multiple MVS datasets show the superior performance of the proposed method. With the same settings in backbone network, our proposed SDA-MVS outperforms its fully-supervised and unsupervised baselines.