论文标题

Vaesim:一种自我监督原型发现的概率方法

VAESim: A probabilistic approach for self-supervised prototype discovery

论文作者

Ferrante, Matteo, Boccato, Tommaso, Spasov, Simeon, Duggento, Andrea, Toschi, Nicola

论文摘要

在医学中,精心策划的图像数据集经常采用离散标签来描述所谓的健康状况与病理状况的连续光谱,例如阿尔茨海默氏病连续体或图像在诊断中起关键点的其他领域。我们建议基于条件变异自动编码器的图像分层的体系结构。我们的框架Vaesim利用连续的潜在空间来表示疾病的连续体并在训练过程中找到簇,然后可以用于图像/患者分层。该方法的核心学习一组原型向量,每个向量与群集相关联。首先,我们将每个数据样本的软分配给群集。然后,我们根据样品嵌入和簇的原型向量之间的相似性度量重建样品。为了更新原型嵌入,我们使用批处理大小中实际原型和样本之间最相似表示的指数移动平均值。我们在MNIST手写数字数据集和名为Pneumoniamnist的医疗基准数据集上测试了我们的方法。我们证明,我们的方法在两个数据集中针对标准VAE的分类任务(性能提高了15%)的KNN精度优于基准,并且还以完全监督的方式培训的分类模型。我们还展示了我们的模型如何优于无监督分层的当前,端到端模型。

In medicine, curated image datasets often employ discrete labels to describe what is known to be a continuous spectrum of healthy to pathological conditions, such as e.g. the Alzheimer's Disease Continuum or other areas where the image plays a pivotal point in diagnosis. We propose an architecture for image stratification based on a conditional variational autoencoder. Our framework, VAESim, leverages a continuous latent space to represent the continuum of disorders and finds clusters during training, which can then be used for image/patient stratification. The core of the method learns a set of prototypical vectors, each associated with a cluster. First, we perform a soft assignment of each data sample to the clusters. Then, we reconstruct the sample based on a similarity measure between the sample embedding and the prototypical vectors of the clusters. To update the prototypical embeddings, we use an exponential moving average of the most similar representations between actual prototypes and samples in the batch size. We test our approach on the MNIST-handwritten digit dataset and on a medical benchmark dataset called PneumoniaMNIST. We demonstrate that our method outperforms baselines in terms of kNN accuracy measured on a classification task against a standard VAE (up to 15% improvement in performance) in both datasets, and also performs at par with classification models trained in a fully supervised way. We also demonstrate how our model outperforms current, end-to-end models for unsupervised stratification.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源