论文标题
元学习,具有上下文不合时宜的初始化
Meta-Learning with Context-Agnostic Initialisations
论文作者
论文摘要
元学习方法通过寻找适合定位任务的初始化来解决的问题。通常,培训数据(我们称为上下文)中有其他属性,与目标任务无关,这些属性是元学习的干扰物,尤其是当目标任务包含训练过程中未见的新颖环境中的示例时。我们通过将上下文 - 对抗组件纳入元学习过程来解决此监督。这会产生一个对目标进行微调的初始化,这既是上下文不可或缺的,又是任务总理。我们评估了三种常用的元学习算法和两个问题的方法。我们证明了我们的上下文不足的元学习可以改善每种情况的结果。首先,我们使用字母作为上下文报告了Omniglot几乎没有弹出字符分类。当从看不见的字母中分类字符时,在方法和任务中的平均提高4.3%。其次,我们以参与者知识为背景在数据集上评估了视频的个性化能源支出预测。我们证明,上下文反应式元学习将平均均方误差降低了30%。
Meta-learning approaches have addressed few-shot problems by finding initialisations suited for fine-tuning to target tasks. Often there are additional properties within training data (which we refer to as context), not relevant to the target task, which act as a distractor to meta-learning, particularly when the target task contains examples from a novel context not seen during training. We address this oversight by incorporating a context-adversarial component into the meta-learning process. This produces an initialisation for fine-tuning to target which is both context-agnostic and task-generalised. We evaluate our approach on three commonly used meta-learning algorithms and two problems. We demonstrate our context-agnostic meta-learning improves results in each case. First, we report on Omniglot few-shot character classification, using alphabets as context. An average improvement of 4.3% is observed across methods and tasks when classifying characters from an unseen alphabet. Second, we evaluate on a dataset for personalised energy expenditure predictions from video, using participant knowledge as context. We demonstrate that context-agnostic meta-learning decreases the average mean square error by 30%.