论文标题
将人解析与分析特征提取和高将来人重新识别的分析特征提取和排名方案
Combining human parsing with analytical feature extraction and ranking schemes for high-generalization person reidentification
论文作者
论文摘要
近年来,由于其对科学和社会的重要性,人们的重新认同(RE-ID)一直受到越来越多的关注。机器学习,尤其是深度学习(DL)已成为主要的重新ID工具,该工具允许研究在基准数据集上达到前所未有的精度水平。但是,DL模型的概括性不佳存在已知的问题。也就是说,经过训练以实现一个数据集的模型在另一个数据集上的表现较差,并且需要重新训练。为了解决这个问题,我们提出了一个没有可训练参数的模型,该模型显示出高概括的巨大潜力。它将完全分析的特征提取和相似性排名方案与用于获得初始子区域分类的基于DL的人解析相结合。我们表明,这种组合在很大程度上消除了现有分析方法的缺点。我们使用可解释的颜色和纹理特征,这些功能具有与之相关的人类可读性相似性措施。为了验证提出的方法,我们在Market1501和CuHK03数据集上进行实验,以达到竞争等级-1精度可与DL模型相当。最重要的是,我们表明,当应用于转移学习任务时,我们的方法可实现63.9%和93.5%的跨域准确性。它显着高于先前报道的30-50%传输精度。我们讨论添加新功能以进一步改善模型的潜在方法。我们还展示了可解释功能的优势,用于构建口头描述中的人类生成的查询,以进行无查询图像进行搜索。
Person reidentification (re-ID) has been receiving increasing attention in recent years due to its importance for both science and society. Machine learning and particularly Deep Learning (DL) has become the main re-id tool that allowed researches to achieve unprecedented accuracy levels on benchmark datasets. However, there is a known problem of poor generalization of DL models. That is, models trained to achieve high accuracy on one dataset perform poorly on other ones and require re-training. To address this issue, we present a model without trainable parameters which shows great potential for high generalization. It combines a fully analytical feature extraction and similarity ranking scheme with DL-based human parsing used to obtain the initial subregion classification. We show that such combination to a high extent eliminates the drawbacks of existing analytical methods. We use interpretable color and texture features which have human-readable similarity measures associated with them. To verify the proposed method we conduct experiments on Market1501 and CUHK03 datasets achieving competitive rank-1 accuracy comparable with that of DL-models. Most importantly we show that our method achieves 63.9% and 93.5% rank-1 cross-domain accuracy when applied to transfer learning tasks. It is significantly higher than previously reported 30-50% transfer accuracy. We discuss the potential ways of adding new features to further improve the model. We also show the advantage of interpretable features for constructing human-generated queries from verbal description to conduct search without a query image.