论文标题
SemifredDonets:部分冷冻的神经网络,用于有效的计算机视觉系统
SemifreddoNets: Partially Frozen Neural Networks for Efficient Computer Vision Systems
论文作者
论文摘要
我们提出了一个系统,该系统由具有部分冷冻权重的固定流程神经网络组成,名为SemifredDonets。 SemifredDonets充当了全面的硬件块,这些硬件块已被优化,具有有效的硬件实现。这些封锁冻结了每一层参数的一定部分,并用固定缩放器替换相应的乘数。固定重量会减少硅区域,逻辑延迟和内存要求,从而大量节省成本和功耗。与传统的层冻结方法不同,半叶型人通过在模型中以不同的尺度和抽象水平配置的一些权重,可以在成本和灵活性之间进行有利可图的交易。尽管修复拓扑结构和某些权重有所限制了灵活性,但我们认为,对于许多用例,该策略的效率优势大于完全可配置模型的优势。此外,我们的系统使用可重复的块,因此它具有调整模型复杂性而无需任何硬件更改的灵活性。与通用加速器上的等效实现相比,半fimefdonets的硬件实现可提供硅面积和功耗的数量级降低。
We propose a system comprised of fixed-topology neural networks having partially frozen weights, named SemifreddoNets. SemifreddoNets work as fully-pipelined hardware blocks that are optimized to have an efficient hardware implementation. Those blocks freeze a certain portion of the parameters at every layer and replace the corresponding multipliers with fixed scalers. Fixing the weights reduces the silicon area, logic delay, and memory requirements, leading to significant savings in cost and power consumption. Unlike traditional layer-wise freezing approaches, SemifreddoNets make a profitable trade between the cost and flexibility by having some of the weights configurable at different scales and levels of abstraction in the model. Although fixing the topology and some of the weights somewhat limits the flexibility, we argue that the efficiency benefits of this strategy outweigh the advantages of a fully configurable model for many use cases. Furthermore, our system uses repeatable blocks, therefore it has the flexibility to adjust model complexity without requiring any hardware change. The hardware implementation of SemifreddoNets provides up to an order of magnitude reduction in silicon area and power consumption as compared to their equivalent implementation on a general-purpose accelerator.