论文标题
基于部分信息分解的神经表示复杂性的度量
A Measure of the Complexity of Neural Representations based on Partial Information Decomposition
论文作者
论文摘要
在神经网络中,与任务相关的信息由神经元组共同表示。但是,对分类标签的共同信息分布在单个神经元之间的特定方式尚不清楚:虽然部分只能从特定的单个神经元中获得部分,但其他部分则由多个神经元进行冗余或协同携带。我们展示了部分信息分解(PID)是信息理论的最新扩展,可以解散这些不同的贡献。由此,我们介绍了“表示复杂性”的度量,该量度量化了访问跨多个神经元信息的难度。我们展示了这种复杂性如何直接适用于较小的层。对于较大的层,我们提出了子采样和粗粒程序,并证明了后者的相应边界。从经验上讲,对于量化MNIST和CIFAR10任务的深度神经网络,我们观察到,代表性复杂性通过连续的隐藏层和过度训练均降低,并将结果与相关措施进行比较。总体而言,我们提出代表性复杂性是一种有原则且可解释的摘要统计量,用于分析神经表示和复杂系统的结构和演变。
In neural networks, task-relevant information is represented jointly by groups of neurons. However, the specific way in which this mutual information about the classification label is distributed among the individual neurons is not well understood: While parts of it may only be obtainable from specific single neurons, other parts are carried redundantly or synergistically by multiple neurons. We show how Partial Information Decomposition (PID), a recent extension of information theory, can disentangle these different contributions. From this, we introduce the measure of "Representational Complexity", which quantifies the difficulty of accessing information spread across multiple neurons. We show how this complexity is directly computable for smaller layers. For larger layers, we propose subsampling and coarse-graining procedures and prove corresponding bounds on the latter. Empirically, for quantized deep neural networks solving the MNIST and CIFAR10 tasks, we observe that representational complexity decreases both through successive hidden layers and over training, and compare the results to related measures. Overall, we propose representational complexity as a principled and interpretable summary statistic for analyzing the structure and evolution of neural representations and complex systems in general.