论文标题
最接近质心的视觉识别
Visual Recognition with Deep Nearest Centroids
论文作者
论文摘要
我们通过重新访问最近的质心,这是最经典,最简单的分类器之一,这是一个在概念上优雅而有效的网络上,它是一个概念上优雅而有效的网络。当前的深层模型以完全参数的方式学习分类器,忽略了潜在的数据结构,缺乏简单性和解释性。 DNC相反进行了非参数,基于案例的推理;它利用培训样本的亚琴师来描述类别分布,并清楚地将分类解释为特征空间中测试数据的接近性和类亚辅助物。由于基于距离的性质,网络输出维度是灵活的,所有可学习的参数仅用于数据嵌入。这意味着,在“预训练和微调”范式下,可以将所有用于像素识别学习的知识完全转移到像素识别学习中。除了其嵌套的简单性和直观的决策机制外,DNC甚至还可以选择次要的解释性,当时将亚中心人选为人类可以查看和检查的实际训练图像。与参数对应物相比,DNC在图像分类(CIFAR-10,IMAGENET)上的表现更好,并且使用各种网络体系结构(Resnet,Swin)和分裂模型(FCN,DeepLabV3,Swinn,Swinn,Swinn,Swinn,Swinn,Swinn,Swin)。我们认为这项工作带来了对相关领域的基本见解。
We devise deep nearest centroids (DNC), a conceptually elegant yet surprisingly effective network for large-scale visual recognition, by revisiting Nearest Centroids, one of the most classic and simple classifiers. Current deep models learn the classifier in a fully parametric manner, ignoring the latent data structure and lacking simplicity and explainability. DNC instead conducts nonparametric, case-based reasoning; it utilizes sub-centroids of training samples to describe class distributions and clearly explains the classification as the proximity of test data and the class sub-centroids in the feature space. Due to the distance-based nature, the network output dimensionality is flexible, and all the learnable parameters are only for data embedding. That means all the knowledge learnt for ImageNet classification can be completely transferred for pixel recognition learning, under the "pre-training and fine-tuning" paradigm. Apart from its nested simplicity and intuitive decision-making mechanism, DNC can even possess ad-hoc explainability when the sub-centroids are selected as actual training images that humans can view and inspect. Compared with parametric counterparts, DNC performs better on image classification (CIFAR-10, ImageNet) and greatly boots pixel recognition (ADE20K, Cityscapes), with improved transparency and fewer learnable parameters, using various network architectures (ResNet, Swin) and segmentation models (FCN, DeepLabV3, Swin). We feel this work brings fundamental insights into related fields.