论文标题
学习使用多尺度对抗性注意门从涂鸦中细分
Learning to Segment from Scribbles using Multi-scale Adversarial Attention Gates
论文作者
论文摘要
在像素级别上注释的大型,细颗粒的图像分割数据集很难获得,尤其是在医学成像中,注释也需要专家知识。弱监督的学习可以通过依靠较弱的注释形式(例如涂鸦)来训练模型。在这里,我们学会在对抗游戏中使用涂鸦注释进行细分。有了未配对的分割掩码,我们训练一个多尺度的gan,以多种分辨率生成现实的分割掩码,而我们使用涂鸦来学习它们在图像中的正确位置。该模型成功的核心是一种新型的注意门控机制,我们用对抗信号来调节该机制,以便先验起形状,从而在多个尺度上进行更好的对象定位。在受对抗调节的前提下,分段者学习了语义的注意图,抑制物体外部的嘈杂激活,并减少分段较深层中消失的梯度问题。我们评估了几种医疗(ACDC,LVSC,混乱)和非医疗(PPSS)数据集的模型,并报告了与经过完全注释的分段掩码训练的模型相匹配的性能水平。我们还展示了各种环境中的扩展:半监督学习;结合多个涂鸦来源(众包场景)和多任务学习(结合涂鸦和面具监督)。我们在https://vios-s.github.io/multiscale-multiscale-versarial-cripting-gates上发布了ACDC数据集的专家制作的涂鸦注释以及实验的代码
Large, fine-grained image segmentation datasets, annotated at pixel-level, are difficult to obtain, particularly in medical imaging, where annotations also require expert knowledge. Weakly-supervised learning can train models by relying on weaker forms of annotation, such as scribbles. Here, we learn to segment using scribble annotations in an adversarial game. With unpaired segmentation masks, we train a multi-scale GAN to generate realistic segmentation masks at multiple resolutions, while we use scribbles to learn their correct position in the image. Central to the model's success is a novel attention gating mechanism, which we condition with adversarial signals to act as a shape prior, resulting in better object localization at multiple scales. Subject to adversarial conditioning, the segmentor learns attention maps that are semantic, suppress the noisy activations outside the objects, and reduce the vanishing gradient problem in the deeper layers of the segmentor. We evaluated our model on several medical (ACDC, LVSC, CHAOS) and non-medical (PPSS) datasets, and we report performance levels matching those achieved by models trained with fully annotated segmentation masks. We also demonstrate extensions in a variety of settings: semi-supervised learning; combining multiple scribble sources (a crowdsourcing scenario) and multi-task learning (combining scribble and mask supervision). We release expert-made scribble annotations for the ACDC dataset, and the code used for the experiments, at https://vios-s.github.io/multiscale-adversarial-attention-gates