论文标题

标签上的损失:通过直接损失构建的弱监督学习

Losses over Labels: Weakly Supervised Learning via Direct Loss Construction

论文作者

Sam, Dylan, Kolter, J. Zico

论文摘要

由于产生大量标记数据的高昂成本,程序化弱监督是机器学习中的一个范式。在这种情况下,用户设计了为数据子集提供嘈杂标签的启发式方法。这些弱标记是组合(通常是通过图形模型)形成伪标记,然后将其用于训练下游模型。在这项工作中,我们质疑典型弱监督的学习管道的基本前提:鉴于启发式启发式提供了所有“标签”信息,我们为什么还需要生成伪标记,而是要直接将启发式方法直接转变为启示范围的损失功能,从而使我们的模型和较高的损失之间的差异更加易于构建。管道,例如启发式方法如何做出决定,在培训期间,我们将其置于训练中的损失(LOL)。

Owing to the prohibitive costs of generating large amounts of labeled data, programmatic weak supervision is a growing paradigm within machine learning. In this setting, users design heuristics that provide noisy labels for subsets of the data. These weak labels are combined (typically via a graphical model) to form pseudolabels, which are then used to train a downstream model. In this work, we question a foundational premise of the typical weakly supervised learning pipeline: given that the heuristic provides all ``label" information, why do we need to generate pseudolabels at all? Instead, we propose to directly transform the heuristics themselves into corresponding loss functions that penalize differences between our model and the heuristic. By constructing losses directly from the heuristics, we can incorporate more information than is used in the standard weakly supervised pipeline, such as how the heuristics make their decisions, which explicitly informs feature selection during training. We call our method Losses over Labels (LoL) as it creates losses directly from heuristics without going through the intermediate step of a label. We show that LoL improves upon existing weak supervision methods on several benchmark text and image classification tasks and further demonstrate that incorporating gradient information leads to better performance on almost every task.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源