论文标题
在低标签制度中重新审视半监督学习
Revisiting Pretraining for Semi-Supervised Learning in the Low-Label Regime
论文作者
论文摘要
半监督学习(SSL)通过通过伪标记来利用大型未标记数据来解决缺乏标记数据的问题。但是,在极低标签的制度中,伪标签可能是不正确的,也就是确认偏差,伪标签反过来又会损害网络培训。最近的研究结合了预处理的权重(FT)和SSL,以减轻挑战,并声称在低标签状态下取得了卓越的结果。在这项工作中,我们首先表明,FT帐户为最先进的表现带来了更好的体重,而且重要的是,他们对现成的半手审查的学习者普遍有帮助。我们进一步认为,由于协变量的偏移,预处理重量的直接填充是次优的,并提出了一个对比目标预训练步骤,以使模型权重适应目标数据集。我们通过进行目标预处理,然后进行半监督的芬太尼,对分类和分割任务进行了广泛的实验。有希望的结果验证了SSL的靶标预处理的功效,特别是在低标签状态下。
Semi-supervised learning (SSL) addresses the lack of labeled data by exploiting large unlabeled data through pseudolabeling. However, in the extremely low-label regime, pseudo labels could be incorrect, a.k.a. the confirmation bias, and the pseudo labels will in turn harm the network training. Recent studies combined finetuning (FT) from pretrained weights with SSL to mitigate the challenges and claimed superior results in the low-label regime. In this work, we first show that the better pretrained weights brought in by FT account for the state-of-the-art performance, and importantly that they are universally helpful to off-the-shelf semi-supervised learners. We further argue that direct finetuning from pretrained weights is suboptimal due to covariate shift and propose a contrastive target pretraining step to adapt model weights towards target dataset. We carried out extensive experiments on both classification and segmentation tasks by doing target pretraining then followed by semi-supervised finetuning. The promising results validate the efficacy of target pretraining for SSL, in particular in the low-label regime.