论文标题
最大征收网络
Maximum-and-Concatenation Networks
论文作者
论文摘要
尽管在许多领域都取得了成功,但深度神经网络(DNN)仍然遭受一些开放问题的困扰,例如本地最小值和不令人满意的概括性能。在这项工作中,我们提出了一种称为最大和征信网络(MCN)的新型体系结构,以尝试消除不良的本地最小值并提高概括能力。值得注意的是,我们证明了MCN拥有非常好的财产。也就是说,\ emph {$(l+1)$ - 层MCN的每个本地最低最低最低限度比至少与包含其第一个$ l $ layers组成的全局最小值一样好。换句话说,通过增加网络深度,MCN可以自主改善其本地最小值的好处,此外,\ emph {很容易将MCN插入现有的深层模型以使其也具有此属性}。最后,在轻度条件下,我们表明MCN可以使用\ emph {高效率}任意近似某些连续功能。也就是说,MCN的覆盖量比大多数现有的DNN(例如Deep Relu)小得多。基于此,我们进一步提供了一个紧密的概括,以确保与测试样品打交道时MCN的推理能力。
While successful in many fields, deep neural networks (DNNs) still suffer from some open problems such as bad local minima and unsatisfactory generalization performance. In this work, we propose a novel architecture called Maximum-and-Concatenation Networks (MCN) to try eliminating bad local minima and improving generalization ability as well. Remarkably, we prove that MCN has a very nice property; that is, \emph{every local minimum of an $(l+1)$-layer MCN can be better than, at least as good as, the global minima of the network consisting of its first $l$ layers}. In other words, by increasing the network depth, MCN can autonomously improve its local minima's goodness, what is more, \emph{it is easy to plug MCN into an existing deep model to make it also have this property}. Finally, under mild conditions, we show that MCN can approximate certain continuous functions arbitrarily well with \emph{high efficiency}; that is, the covering number of MCN is much smaller than most existing DNNs such as deep ReLU. Based on this, we further provide a tight generalization bound to guarantee the inference ability of MCN when dealing with testing samples.