论文标题

困难网络:学习预测长尾识别的困难

Difficulty-Net: Learning to Predict Difficulty for Long-Tailed Recognition

论文作者

Sinha, Saptarshi, Ohashi, Hiroki

论文摘要

长尾数据集(Head Class)组成的培训样本比尾巴类别多得多,这会导致识别模型对头等阶层有偏见。加权损失是缓解此问题的最流行方法之一,最近的一项工作表明,与常规使用的班级频率相比,班级难度可能是一个更好的线索。在先前的工作中使用了一种启发式公式来量化难度,但是我们从经验上发现,最佳配方取决于数据集的特征。因此,我们提出了困难网络,该难题学习在元学习框架中使用模型的性能来预测类的难度。为了使其在其他班级的背景下学习班级的合理难度,我们新介绍了两个关键概念,即相对困难和驾驶员损失。在计算班级的难度时,前者有助于难以确定其他课程,而后者对于将学习指向有意义的方向是必不可少的。对流行的长尾数据集进行了广泛的实验证明了该方法的有效性,并且在多个长尾数据集上实现了最先进的性能。

Long-tailed datasets, where head classes comprise much more training samples than tail classes, cause recognition models to get biased towards the head classes. Weighted loss is one of the most popular ways of mitigating this issue, and a recent work has suggested that class-difficulty might be a better clue than conventionally used class-frequency to decide the distribution of weights. A heuristic formulation was used in the previous work for quantifying the difficulty, but we empirically find that the optimal formulation varies depending on the characteristics of datasets. Therefore, we propose Difficulty-Net, which learns to predict the difficulty of classes using the model's performance in a meta-learning framework. To make it learn reasonable difficulty of a class within the context of other classes, we newly introduce two key concepts, namely the relative difficulty and the driver loss. The former helps Difficulty-Net take other classes into account when calculating difficulty of a class, while the latter is indispensable for guiding the learning to a meaningful direction. Extensive experiments on popular long-tailed datasets demonstrated the effectiveness of the proposed method, and it achieved state-of-the-art performance on multiple long-tailed datasets.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源