论文标题

任何弹出的对象检测

Any-Shot Object Detection

论文作者

Rahman, Shafin, Khan, Salman, Barnes, Nick, Khan, Fahad Shahbaz

论文摘要

关于新颖对象检测的先前工作考虑了零或几次射击设置,在这些设置中,每个类别的示例都没有或几个示例可供培训。在现实世界的情况下,预期“所有”新颖的班级是看不见的,或者{{有}很少的审查是不太实际的。在这里,我们提出了一个更现实的设置,称为“任何射击检测”,在推理过程中,完全看不见和几乎没有射击类别可以同时共发生。与传统的新颖对象检测相比,任何镜头检测都会带来独特的挑战,例如,在学习新颖的课程时忘记基础训练并将新颖的课程与背景区分开来,忘记基础训练的对象类别之间的高度失衡,忘记了基础训练。为了应对这些挑战,我们提出了一个统一的任何照片检测模型,可以同时学习检测零射击和少量射击对象类。我们的核心思想是将类语义用作对象检测的原型,该公式自然可以最大程度地减少知识遗忘并减轻标签空间中的类不平衡。此外,我们提出了一种重新平衡的损失功能,强调了很难少的案例,但避免了新颖的类别过度拟合,以允许发现完全看不见的类别。如果没有铃铛和口哨声,我们的框架也可以仅用于零弹检测和少量检测任务。我们报告了有关Pascal VOC和MS-Coco数据集的广泛实验,我们的方法被证明可以提供重大改进。

Previous work on novel object detection considers zero or few-shot settings where none or few examples of each category are available for training. In real world scenarios, it is less practical to expect that 'all' the novel classes are either unseen or {have} few-examples. Here, we propose a more realistic setting termed 'Any-shot detection', where totally unseen and few-shot categories can simultaneously co-occur during inference. Any-shot detection offers unique challenges compared to conventional novel object detection such as, a high imbalance between unseen, few-shot and seen object classes, susceptibility to forget base-training while learning novel classes and distinguishing novel classes from the background. To address these challenges, we propose a unified any-shot detection model, that can concurrently learn to detect both zero-shot and few-shot object classes. Our core idea is to use class semantics as prototypes for object detection, a formulation that naturally minimizes knowledge forgetting and mitigates the class-imbalance in the label space. Besides, we propose a rebalanced loss function that emphasizes difficult few-shot cases but avoids overfitting on the novel classes to allow detection of totally unseen classes. Without bells and whistles, our framework can also be used solely for Zero-shot detection and Few-shot detection tasks. We report extensive experiments on Pascal VOC and MS-COCO datasets where our approach is shown to provide significant improvements.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源