论文标题
通过有效的培训策略加速自我监督学习
Accelerating Self-Supervised Learning via Efficient Training Strategies
论文作者
论文摘要
最近,计算机视觉社区的重点已从昂贵的监督学习转变为对视觉表示的自我监督学习。尽管受监督和自我监督之间的性能差距正在缩小,但培训自我监督的深层网络的时间仍然比其监督同行的数量级要大,这阻碍了进步,施加碳成本,并限制了社会利益,将社会福利限制为机构对机构的大量资源。在这些问题的推动下,本文调查了尚未用于此问题的各种模型 - 敏捷策略减少了最近的自我监督方法的训练时间。特别是,我们研究了三种策略:可扩展的循环学习率计划,匹配的渐进式增强幅度和图像分辨率计划以及基于增强难度的硬式挖掘策略。我们表明,在几种自我监管方法的训练时间内,这三种方法的结合速度高达2.7倍,同时保留了与标准的自我监督学习设置相当的性能。
Recently the focus of the computer vision community has shifted from expensive supervised learning towards self-supervised learning of visual representations. While the performance gap between supervised and self-supervised has been narrowing, the time for training self-supervised deep networks remains an order of magnitude larger than its supervised counterparts, which hinders progress, imposes carbon cost, and limits societal benefits to institutions with substantial resources. Motivated by these issues, this paper investigates reducing the training time of recent self-supervised methods by various model-agnostic strategies that have not been used for this problem. In particular, we study three strategies: an extendable cyclic learning rate schedule, a matching progressive augmentation magnitude and image resolutions schedule, and a hard positive mining strategy based on augmentation difficulty. We show that all three methods combined lead up to 2.7 times speed-up in the training time of several self-supervised methods while retaining comparable performance to the standard self-supervised learning setting.