论文标题

深度光谱方法:无监督语义分割和本地化的令人惊讶的强大基线

Deep Spectral Methods: A Surprisingly Strong Baseline for Unsupervised Semantic Segmentation and Localization

论文作者

Melas-Kyriazi, Luke, Rupprecht, Christian, Laina, Iro, Vedaldi, Andrea

论文摘要

无监督的本地化和细分是长期以来的计算机视觉挑战,涉及将图像分解为语义上的片段,而没有任何标记的数据。由于获得密集图像注释的困难和成本,这些任务在无监督的环境中尤其有趣,但是现有的无监督方法与包含多个对象的复杂场景困难。与现有方法纯粹基于深度学习的方式不同,我们通过将图像分解作为图形分配问题来汲取传统光谱分割方法的灵感。具体而言,我们检查了来自自我监管的网络的特征亲和力矩阵的拉普拉斯的特征向量。我们发现这些特征向量已经将图像分解为有意义的片段,并且可以轻易地用于在场景中定位对象。此外,通过将与这些段相关联的特征在数据集中聚集,我们可以获得精心缩减的,可命名的区域,即语义分割。在复杂数据集(Pascal VOC,MS-Coco)上进行的实验表明,我们的简单光谱方法优于无监督的本地化和分割的最先进的方法。此外,我们的方法可以很容易地用于各种复杂的图像编辑任务,例如背景删除和合成。

Unsupervised localization and segmentation are long-standing computer vision challenges that involve decomposing an image into semantically-meaningful segments without any labeled data. These tasks are particularly interesting in an unsupervised setting due to the difficulty and cost of obtaining dense image annotations, but existing unsupervised approaches struggle with complex scenes containing multiple objects. Differently from existing methods, which are purely based on deep learning, we take inspiration from traditional spectral segmentation methods by reframing image decomposition as a graph partitioning problem. Specifically, we examine the eigenvectors of the Laplacian of a feature affinity matrix from self-supervised networks. We find that these eigenvectors already decompose an image into meaningful segments, and can be readily used to localize objects in a scene. Furthermore, by clustering the features associated with these segments across a dataset, we can obtain well-delineated, nameable regions, i.e. semantic segmentations. Experiments on complex datasets (Pascal VOC, MS-COCO) demonstrate that our simple spectral method outperforms the state-of-the-art in unsupervised localization and segmentation by a significant margin. Furthermore, our method can be readily used for a variety of complex image editing tasks, such as background removal and compositing.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源