论文标题
有条件的负面抽样,用于对比度学习视觉表示
Conditional Negative Sampling for Contrastive Learning of Visual Representations
论文作者
论文摘要
学习无监督的视觉表示的最新方法,称为对比度学习,优化了图像两种观点之间的相互信息绑定的噪声对抗性估计(NCE)。 NCE使用随机采样的负示例来正常化目标。在本文中,我们表明,选择困难的负面因素或与当前实例更相似的否定性可以产生更强的表示。为此,我们介绍了一个共同信息估算器的家族,该家族有条件地采样否定的,以每个正面的“戒指”在“环”中。我们证明,这些估计量较低的互信息,偏差较高,但差异较低。在实验上,我们发现我们的方法应用于现有模型(IR,CMC和MOCO)的顶部,在每种情况下,将精度提高了2-5%,通过在四个标准图像数据集上进行线性评估来衡量。此外,我们发现将功能转移到从元数据集合和各种下游任务(例如对象检测,实例细分和关键点检测)等各种下游任务的新图像分布中时,我们发现了持续的好处。
Recent methods for learning unsupervised visual representations, dubbed contrastive learning, optimize the noise-contrastive estimation (NCE) bound on mutual information between two views of an image. NCE uses randomly sampled negative examples to normalize the objective. In this paper, we show that choosing difficult negatives, or those more similar to the current instance, can yield stronger representations. To do this, we introduce a family of mutual information estimators that sample negatives conditionally -- in a "ring" around each positive. We prove that these estimators lower-bound mutual information, with higher bias but lower variance than NCE. Experimentally, we find our approach, applied on top of existing models (IR, CMC, and MoCo) improves accuracy by 2-5% points in each case, measured by linear evaluation on four standard image datasets. Moreover, we find continued benefits when transferring features to a variety of new image distributions from the Meta-Dataset collection and to a variety of downstream tasks such as object detection, instance segmentation, and keypoint detection.