论文标题
Nevis'22:从30年的计算机视觉研究中采样了100个任务流
NEVIS'22: A Stream of 100 Tasks Sampled from 30 Years of Computer Vision Research
论文作者
论文摘要
诸如持续学习,元学习和转移学习之类的几个机器学习社区的共同目标是设计有效,稳健地适应看不见的任务的算法和模型。一个更加雄心勃勃的目标是建立永不停止适应的模型,并且通过适当地转移应计知识而变得越来越有效。除了研究实际学习算法和模型体系结构外,还有一些障碍,我们要建立此类模型,例如学习协议的选择,成功指标以及验证研究假设所需的数据。在这项工作中,我们介绍了永无止境的视觉分类流(Nevis'22),这是一个由100多个视觉分类任务的基准组成的基准,按时间顺序排序,并从从过去三十年的计算机视觉程序中均匀采样的论文中提取。最终的流反映了研究社区在任何时间点的有意义,它是一个理想的测试床,可以评估模型能够适应新任务的能力,并且随着时间的流逝而做得更好,更有效。尽管仅限于分类,但最终的流从OCR到纹理分析,场景识别等各种各样的任务。多样性也反映在各种数据集尺寸中,超过四个数量级。总体而言,由于任务的规模和多样性,Nevis'22对当前的顺序学习方法提出了前所未有的挑战,但由于障碍的范围较低,因此它仅限于单一模态和充分理解的监督学习问题。此外,我们提供了参考实施,包括强大的基准和评估协议,以在精度和计算之间的权衡方面比较方法。
A shared goal of several machine learning communities like continual learning, meta-learning and transfer learning, is to design algorithms and models that efficiently and robustly adapt to unseen tasks. An even more ambitious goal is to build models that never stop adapting, and that become increasingly more efficient through time by suitably transferring the accrued knowledge. Beyond the study of the actual learning algorithm and model architecture, there are several hurdles towards our quest to build such models, such as the choice of learning protocol, metric of success and data needed to validate research hypotheses. In this work, we introduce the Never-Ending VIsual-classification Stream (NEVIS'22), a benchmark consisting of a stream of over 100 visual classification tasks, sorted chronologically and extracted from papers sampled uniformly from computer vision proceedings spanning the last three decades. The resulting stream reflects what the research community thought was meaningful at any point in time, and it serves as an ideal test bed to assess how well models can adapt to new tasks, and do so better and more efficiently as time goes by. Despite being limited to classification, the resulting stream has a rich diversity of tasks from OCR, to texture analysis, scene recognition, and so forth. The diversity is also reflected in the wide range of dataset sizes, spanning over four orders of magnitude. Overall, NEVIS'22 poses an unprecedented challenge for current sequential learning approaches due to the scale and diversity of tasks, yet with a low entry barrier as it is limited to a single modality and well understood supervised learning problems. Moreover, we provide a reference implementation including strong baselines and an evaluation protocol to compare methods in terms of their trade-off between accuracy and compute.