论文标题

对比度学习深度强大的聚类

Deep Robust Clustering by Contrastive Learning

论文作者

Zhong, Huasong, Chen, Chong, Jin, Zhongming, Hua, Xian-Sheng

论文摘要

最近,已经提出了许多无监督的深度学习方法,以通过未标记的数据学习聚类。通过引入数据增强,大多数最新方法从原始图像及其转换应共享相似的语义聚类分配的角度来看深度聚类。但是,由于SoftMax函数仅对最大值敏感,即使将它们分配到同一群集也可能完全不同。这可能会导致代表特征空间中较高的类内多样性,这将导致局部最佳最佳,从而损害聚类性能。为了解决这个缺点,我们提出了深度鲁棒的聚类(DRC)。与现有方法不同,刚果民主共和国从语义聚类分配和表示特征的两个角度进行了深入的聚类,它们可以同时增加阶层间多样性并减少阶层内多样性。此外,我们总结了一个通用框架,该框架可以通过研究相互信息和对比度学习之间的内部关系,将任何最大化的相互信息转化为最大程度地减少对比度损失。我们成功地将其应用于DRC,以学习不变的功能和稳健的簇。在六个广泛的深层聚类基准上进行了广泛的实验,证明了DRC在稳定性和准确性方面的优越性。例如,在CIFAR-10上达到71.6%的平均准确性,比最新结果高7.1%。

Recently, many unsupervised deep learning methods have been proposed to learn clustering with unlabelled data. By introducing data augmentation, most of the latest methods look into deep clustering from the perspective that the original image and its transformation should share similar semantic clustering assignment. However, the representation features could be quite different even they are assigned to the same cluster since softmax function is only sensitive to the maximum value. This may result in high intra-class diversities in the representation feature space, which will lead to unstable local optimal and thus harm the clustering performance. To address this drawback, we proposed Deep Robust Clustering (DRC). Different from existing methods, DRC looks into deep clustering from two perspectives of both semantic clustering assignment and representation feature, which can increase inter-class diversities and decrease intra-class diversities simultaneously. Furthermore, we summarized a general framework that can turn any maximizing mutual information into minimizing contrastive loss by investigating the internal relationship between mutual information and contrastive learning. And we successfully applied it in DRC to learn invariant features and robust clusters. Extensive experiments on six widely-adopted deep clustering benchmarks demonstrate the superiority of DRC in both stability and accuracy. e.g., attaining 71.6% mean accuracy on CIFAR-10, which is 7.1% higher than state-of-the-art results.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源