论文标题

用于插值的前馈relu网络的通用解决方案

Universal Solutions of Feedforward ReLU Networks for Interpolations

论文作者

Huang, Changcun

论文摘要

本文在所谓的插值矩阵方面提供了一个理论框架,用于插值的馈电回路网络的解决方案,这是我们三项上一件作品的摘要,扩展和概括,并期望工程解决方案可以包括在此框架中并最终理解。对于三层网络,我们对不同种类的解决方案进行了分类,并以归一化形式对其进行建模。解决方案发现由三个维度(包括数据,网络和培训)进行研究;解释了一种过度参数化解决方案的机制。对于深层网络,我们提出了一个称为稀疏矩阵原理的一般结果,可以描述深层的某些基本行为,并解释与脑科学相关的工程应用中出现的稀疏激活模式的现象;与较浅的层相比,深层的优势在此原理中表现出来。作为应用,该原则构建了深层神经网络进行分类的一般解决方案。我们还使用该原理来研究编码器的数据访问性属性。类似于三层情况,还通过几个维度探索了深层的溶液。从插值矩阵的角度解释了多输出神经网络的机制。

This paper provides a theoretical framework on the solution of feedforward ReLU networks for interpolations, in terms of what is called an interpolation matrix, which is the summary, extension and generalization of our three preceding works, with the expectation that the solution of engineering could be included in this framework and finally understood. To three-layer networks, we classify different kinds of solutions and model them in a normalized form; the solution finding is investigated by three dimensions, including data, networks and the training; the mechanism of a type of overparameterization solution is interpreted. To deep-layer networks, we present a general result called sparse-matrix principle, which could describe some basic behavior of deep layers and explain the phenomenon of the sparse-activation mode that appears in engineering applications associated with brain science; an advantage of deep layers compared to shallower ones is manifested in this principle. As applications, a general solution of deep neural networks for classifications is constructed by that principle; and we also use the principle to study the data-disentangling property of encoders. Analogous to the three-layer case, the solution of deep layers is also explored through several dimensions. The mechanism of multi-output neural networks is explained from the perspective of interpolation matrices.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源