论文标题

SELECMIX:通过矛盾的对样本抽样的学习

SelecMix: Debiased Learning by Contradicting-pair Sampling

论文作者

Hwang, Inwoo, Lee, Sangjun, Kwak, Yunhyeok, Oh, Seong Joon, Teney, Damien, Kim, Jin-Hwa, Zhang, Byoung-Tak

论文摘要

接受ERM培训的神经网络(经验风险最小化)有时会学习意想不到的决策规则,特别是当他们的培训数据有偏见时,即训练标签与不良功能密切相关时。为了防止网络学习此类功能,最近的方法增加了培训数据,以使显示虚假相关性的示例(即与偏见的示例)成为少数群体,而另一个偏见的示例则普遍存在。但是,这些方法有时很难训练和扩展到现实世界数据,因为它们依赖于生成模型或分离的表示。我们提出了一种基于混音的替代方案,这是一种流行的增强,可以创建培训示例的凸组合。我们的方法(造型的selecmix)将混音应用于矛盾的示例对,定义为显示(i)相同的标签但不同的偏见特征,或(ii)不同的标签,但具有相似的偏见特征。识别此类对需要比较有关未知偏见特征的示例。为此,我们利用了一个流行的启发式方法,即在训练过程中优先学习了偏见的特征。对标准基准测试的实验证明了该方法的有效性,特别是当标记噪声使偏见冲突示例的识别复杂化时。

Neural networks trained with ERM (empirical risk minimization) sometimes learn unintended decision rules, in particular when their training data is biased, i.e., when training labels are strongly correlated with undesirable features. To prevent a network from learning such features, recent methods augment training data such that examples displaying spurious correlations (i.e., bias-aligned examples) become a minority, whereas the other, bias-conflicting examples become prevalent. However, these approaches are sometimes difficult to train and scale to real-world data because they rely on generative models or disentangled representations. We propose an alternative based on mixup, a popular augmentation that creates convex combinations of training examples. Our method, coined SelecMix, applies mixup to contradicting pairs of examples, defined as showing either (i) the same label but dissimilar biased features, or (ii) different labels but similar biased features. Identifying such pairs requires comparing examples with respect to unknown biased features. For this, we utilize an auxiliary contrastive model with the popular heuristic that biased features are learned preferentially during training. Experiments on standard benchmarks demonstrate the effectiveness of the method, in particular when label noise complicates the identification of bias-conflicting examples.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源