论文标题
深度为非本地模块,用于使用单个线程的快速显着对象检测
Depthwise Non-local Module for Fast Salient Object Detection Using a Single Thread
论文作者
论文摘要
最近,深度卷积神经网络在显着对象检测方面取得了重大成功。但是,现有的最新方法需要高端GPU才能实现实时性能,这使它们难以适应低成本或便携式设备。尽管已经提出了通用的网络体系结构来加快移动设备的推断,但它们是根据图像分类或语义细分的任务量身定制的,并难以捕获渠道内和通道间的相关性,这些相关性对于显着对象检测对于对比度建模至关重要。在上述观察中,我们为快速显着对象检测设计了一种新的深度学习算法。提出的算法首次与单个CPU线程同时达到竞争精度和高推理效率。具体而言,我们提出了一种新型的深度非本地粘膜(DNL),它隐含地模型通过以自我注意力方式收集通道内和通道间相关性进行对比。此外,我们引入了深度的非本地网络结构,该结构既包含了深度的非本地模块和倒残留的块。实验结果表明,我们提出的网络在广泛的显着对象检测数据集上具有非常有竞争力的准确性,同时在所有现有的基于深度学习的算法中都达到了最先进的效率。
Recently deep convolutional neural networks have achieved significant success in salient object detection. However, existing state-of-the-art methods require high-end GPUs to achieve real-time performance, which makes them hard to adapt to low-cost or portable devices. Although generic network architectures have been proposed to speed up inference on mobile devices, they are tailored to the task of image classification or semantic segmentation, and struggle to capture intra-channel and inter-channel correlations that are essential for contrast modeling in salient object detection. Motivated by the above observations, we design a new deep learning algorithm for fast salient object detection. The proposed algorithm for the first time achieves competitive accuracy and high inference efficiency simultaneously with a single CPU thread. Specifically, we propose a novel depthwise non-local moudule (DNL), which implicitly models contrast via harvesting intra-channel and inter-channel correlations in a self-attention manner. In addition, we introduce a depthwise non-local network architecture that incorporates both depthwise non-local modules and inverted residual blocks. Experimental results show that our proposed network attains very competitive accuracy on a wide range of salient object detection datasets while achieving state-of-the-art efficiency among all existing deep learning based algorithms.