论文标题

结合自发动模块和特征金字塔网络的显着对象检测

Salient Object Detection Combining a Self-attention Module and a Feature Pyramid Network

论文作者

Ren, Guangyu, Dai, Tianhong, Barmpoutis, Panagiotis, Stathaki, Tania

论文摘要

通过使用完全卷积网络(FCN),显着对象检测取得了重大改进。但是,基于FCN的U形架构可能会在自上而下的途径的上样本操作期间在高级语义信息中引起稀释问题。因此,它可以削弱显着对象定位的能力并产生降级边界。为此,为了克服这一局限性,我们提出了一种新型的金字塔自我发场模块(PSAM),并采用了独立的特征汇编策略。在PSAM中,在多尺度金字塔特征以捕获更丰富的高级特征并为模型带来更大的接收场之后,自我发项层配备了。此外,还采用了渠道注意模块来减少FPN的冗余特征并提供精制结果。实验分析表明,所提出的PSAM有效地有助于整个模型,因此在五个具有挑战性的数据集中优于最先进的结果。最后,定量结果表明,PSAM生成了清晰且积分的显着图,可以为其他计算机视觉任务(例如对象检测和语义分割)提供进一步的帮助。

Salient object detection has achieved great improvement by using the Fully Convolution Network (FCN). However, the FCN-based U-shape architecture may cause the dilution problem in the high-level semantic information during the up-sample operations in the top-down pathway. Thus, it can weaken the ability of salient object localization and produce degraded boundaries. To this end, in order to overcome this limitation, we propose a novel pyramid self-attention module (PSAM) and the adoption of an independent feature-complementing strategy. In PSAM, self-attention layers are equipped after multi-scale pyramid features to capture richer high-level features and bring larger receptive fields to the model. In addition, a channel-wise attention module is also employed to reduce the redundant features of the FPN and provide refined results. Experimental analysis shows that the proposed PSAM effectively contributes to the whole model so that it outperforms state-of-the-art results over five challenging datasets. Finally, quantitative results show that PSAM generates clear and integral salient maps which can provide further help to other computer vision tasks, such as object detection and semantic segmentation.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源