论文标题
用图表代表深神经网络潜在的几何形状
Representing Deep Neural Networks Latent Space Geometries with Graphs
论文作者
论文摘要
深度学习(DL)因其在许多机器学习任务中达到最先进的表现的能力而引起了很多关注。 DL方法的核心原理在于以端到端方式训练复合体系结构,其中输入与经过训练以优化目标函数的输出相关联。由于其组成性质,DL体系结构自然表现出属于所谓潜在空间的输入的几种中间表示。当单独处理时,这些中间表示在学习过程中大部分时间都不受限制,因为尚不清楚应该有利于哪些属性。但是,当同时处理一批输入时,相应的中间表示集显示了可以在其中寻求所需属性的关系(我们称之为几何形状)。在这项工作中,我们表明可以对这些潜在几何形状引入约束以解决各种问题。在更多详细信息中,我们建议通过从处理一批输入时从中间表示中构造相似性图来表示几何图。通过约束这些潜在的几何图(LGG),我们解决了以下三个问题:i)通过模仿其几何形状来实现教师体系结构的行为,ii)ii)设计有效的分类嵌入,通过靶向特定的几何形状来实现特定的几何形式,ii)通过在potitions上进行稳健的范围,从而实现了稳固的效果,从而实现了良好的效果。使用标准视觉基准测试,我们证明了提出的基于几何方法解决所考虑问题的能力。
Deep Learning (DL) has attracted a lot of attention for its ability to reach state-of-the-art performance in many machine learning tasks. The core principle of DL methods consists in training composite architectures in an end-to-end fashion, where inputs are associated with outputs trained to optimize an objective function. Because of their compositional nature, DL architectures naturally exhibit several intermediate representations of the inputs, which belong to so-called latent spaces. When treated individually, these intermediate representations are most of the time unconstrained during the learning process, as it is unclear which properties should be favored. However, when processing a batch of inputs concurrently, the corresponding set of intermediate representations exhibit relations (what we call a geometry) on which desired properties can be sought. In this work, we show that it is possible to introduce constraints on these latent geometries to address various problems. In more details, we propose to represent geometries by constructing similarity graphs from the intermediate representations obtained when processing a batch of inputs. By constraining these Latent Geometry Graphs (LGGs), we address the three following problems: i) Reproducing the behavior of a teacher architecture is achieved by mimicking its geometry, ii) Designing efficient embeddings for classification is achieved by targeting specific geometries, and iii) Robustness to deviations on inputs is achieved via enforcing smooth variation of geometry between consecutive latent spaces. Using standard vision benchmarks, we demonstrate the ability of the proposed geometry-based methods in solving the considered problems.