论文标题

通过加权损失,类不平衡的互补标签学习

Class-Imbalanced Complementary-Label Learning via Weighted Loss

论文作者

Wei, Meng, Zhou, Yong, Li, Zhongnian, Xu, Xinzheng

论文摘要

互补标签的学习(CLL)被广泛用于弱监督分类,但是在面对类造型培训样本时,它在现实世界数据集中面临重大挑战。在这种情况下,一个类别中的样本数量大大低于其他类别,从而导致预测准确性下降。不幸的是,现有的CLL方法尚未调查此问题。为了减轻这一挑战,我们提出了一个新颖的问题设置,该设置可以从平衡的互补标签中学习以进行多类分类。为了解决这个问题,我们提出了一种新型的CLL方法,称为加权互补标签学习(WCLL)。提出的方法通过使用平衡的互补标签来模拟加权的经验风险最小化损失,该标签也适用于多类不平衡训练样本。此外,我们得出一个估计误差,必将提供理论保证。为了评估我们的方法,我们对几个广泛使用的基准数据集和一个现实世界数据集进行了广泛的实验,并将我们的方法与现有的最新方法进行比较。即使在多种类不平衡的情况下,提出的方法也显示出这些数据集的显着改善。值得注意的是,所提出的方法不仅利用互补标签来训练分类器,而且还解决了类不平衡的问题。

Complementary-label learning (CLL) is widely used in weakly supervised classification, but it faces a significant challenge in real-world datasets when confronted with class-imbalanced training samples. In such scenarios, the number of samples in one class is considerably lower than in other classes, which consequently leads to a decline in the accuracy of predictions. Unfortunately, existing CLL approaches have not investigate this problem. To alleviate this challenge, we propose a novel problem setting that enables learning from class-imbalanced complementary labels for multi-class classification. To tackle this problem, we propose a novel CLL approach called Weighted Complementary-Label Learning (WCLL). The proposed method models a weighted empirical risk minimization loss by utilizing the class-imbalanced complementary labels, which is also applicable to multi-class imbalanced training samples. Furthermore, we derive an estimation error bound to provide theoretical assurance. To evaluate our approach, we conduct extensive experiments on several widely-used benchmark datasets and a real-world dataset, and compare our method with existing state-of-the-art methods. The proposed approach shows significant improvement in these datasets, even in the case of multiple class-imbalanced scenarios. Notably, the proposed method not only utilizes complementary labels to train a classifier but also solves the problem of class imbalance.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源