论文标题
广义ODIN:检测到分布图像的图像,而无需从分布数据中学习
Generalized ODIN: Detecting Out-of-distribution Image without Learning from Out-of-distribution Data
论文作者
论文摘要
当应用于与训练集相同的分布的数据时,深层神经网络的性能已经出色,但否则可能会大大降级。因此,检测示例是否不合格(OOD)对于启用可以拒绝此类示例或警报用户的系统至关重要。最近的工作在由小型图像数据集组成的OOD基准上取得了重大进展。但是,许多基于神经网络的最新方法都依赖于分布和分布数据的培训或调整。后者通常很难定义A-Priori,其选择可以轻松偏见学习。我们的工作基于流行的方法ODIN,提出了两种将其从使用OOD数据调整的需求的策略,同时改善了其OOD检测性能。我们特别建议分解置信度评分以及修改的输入预处理方法。我们表明,这两种都极大地有助于检测性能。我们对较大规模图像数据集的进一步分析表明,两种类型的分布变化,特别是语义转移和非语义转移,对问题的难度呈现出显着差异,从而分析了类似Odin的策略何时或不起作用。
Deep neural networks have attained remarkable performance when applied to data that comes from the same distribution as that of the training set, but can significantly degrade otherwise. Therefore, detecting whether an example is out-of-distribution (OoD) is crucial to enable a system that can reject such samples or alert users. Recent works have made significant progress on OoD benchmarks consisting of small image datasets. However, many recent methods based on neural networks rely on training or tuning with both in-distribution and out-of-distribution data. The latter is generally hard to define a-priori, and its selection can easily bias the learning. We base our work on a popular method ODIN, proposing two strategies for freeing it from the needs of tuning with OoD data, while improving its OoD detection performance. We specifically propose to decompose confidence scoring as well as a modified input pre-processing method. We show that both of these significantly help in detection performance. Our further analysis on a larger scale image dataset shows that the two types of distribution shifts, specifically semantic shift and non-semantic shift, present a significant difference in the difficulty of the problem, providing an analysis of when ODIN-like strategies do or do not work.