论文标题
通过生成神经符号建模学习任务代表
Learning Task-General Representations with Generative Neuro-Symbolic Modeling
论文作者
论文摘要
人们只能从原始的感知输入中学习丰富的通用概念表示。当前的机器学习方法远远远远没有达到这些人类标准,尽管不同的建模传统通常具有互补的优势。符号模型可以捕获能够灵活概括的组成和因果知识,但是他们努力从原始输入中学习,依靠强大的抽象和简化假设。神经网络模型可以直接从原始数据中学习,但是它们很难捕获构图和因果结构,通常必须重新培训以解决新任务。我们将这两种传统汇总在一起,以学习概念的生成模型,以捕获丰富的组成和因果结构,同时从原始数据中学习。我们开发了手写字符概念的生成神经符号(GNS)模型,该模型使用了概率程序的控制流,并与符号卒中原语和符号图像渲染器相结合,以代表因果和组成过程,形成了特征的因果和组成过程。零件(中风)的分布以及零件之间的相关性是通过神经网络子例程建模的,允许该模型直接从原始数据中学习并表达非参数统计关系。我们将模型应用于人类水平概念学习的综合挑战,使用一组字母集来学习对字符图的表现力先验分布。在随后的评估中,我们的GNS模型使用概率推断从单个培训图像中学习丰富的概念表示,该图像将其推广到4个独特的任务,在先前工作不足的情况下成功。
People can learn rich, general-purpose conceptual representations from only raw perceptual inputs. Current machine learning approaches fall well short of these human standards, although different modeling traditions often have complementary strengths. Symbolic models can capture the compositional and causal knowledge that enables flexible generalization, but they struggle to learn from raw inputs, relying on strong abstractions and simplifying assumptions. Neural network models can learn directly from raw data, but they struggle to capture compositional and causal structure and typically must retrain to tackle new tasks. We bring together these two traditions to learn generative models of concepts that capture rich compositional and causal structure, while learning from raw data. We develop a generative neuro-symbolic (GNS) model of handwritten character concepts that uses the control flow of a probabilistic program, coupled with symbolic stroke primitives and a symbolic image renderer, to represent the causal and compositional processes by which characters are formed. The distributions of parts (strokes), and correlations between parts, are modeled with neural network subroutines, allowing the model to learn directly from raw data and express nonparametric statistical relationships. We apply our model to the Omniglot challenge of human-level concept learning, using a background set of alphabets to learn an expressive prior distribution over character drawings. In a subsequent evaluation, our GNS model uses probabilistic inference to learn rich conceptual representations from a single training image that generalize to 4 unique tasks, succeeding where previous work has fallen short.